Tropical forests are significant carbon sinks and their soils’ carbon storage potential is immense. However, little is known about the soil organic carbon (SOC) stocks of tropical mountain areas whose complex soil-landscape and difficult accessibility pose a challenge to spatial analysis. The choice of methodology for spatial prediction is of high importance to improve the expected poor model results in case of low predictor-response correlations. Four aspects were considered to improve model performance in predicting SOC stocks of the organic layer of a tropical mountain forest landscape: Different spatial predictor settings, predictor selection strategies, various machine learning algorithms and model tuning. Five machine learning algorithms: random forests, artificial neural networks, multivariate adaptive regression splines, boosted regression trees and support vector machines were trained and tuned to predict SOC stocks from predictors derived from a digital elevation model and satellite image. Topographical predictors were calculated with a GIS search radius of 45 to 615 m. Finally, three predictor selection strategies were applied to the total set of 236 predictors. All machine learning algorithms—including the model tuning and predictor selection—were compared via five repetitions of a tenfold cross-validation. The boosted regression tree algorithm resulted in the overall best model. SOC stocks ranged between 0.2 to 17.7 kg m-2, displaying a huge variability with diffuse insolation and curvatures of different scale guiding the spatial pattern. Predictor selection and model tuning improved the models’ predictive performance in all five machine learning algorithms. The rather low number of selected predictors favours forward compared to backward selection procedures. Choosing predictors due to their indiviual performance was vanquished by the two procedures which accounted for predictor interaction.
Citation: Ließ M, Schmidt J, Glaser B (2016) Improving the Spatial Prediction of Soil Organic Carbon Stocks in a Complex Tropical Mountain Landscape by Methodological Specifications in Machine Learning Approaches. PLoS ONE 11(4): e0153673. https://doi.org/10.1371/journal.pone.0153673
Editor: Mansour Ebrahimi, Qom University, ISLAMIC REPUBLIC OF IRAN
Received: October 23, 2015; Accepted: April 1, 2016; Published: April 29, 2016
Copyright: © 2016 Ließ et al. This is an open access article distributed under the terms of the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited.
Data Availability: Data are uploaded to Dryad under the following DOI: http://dx.doi.org/10.5061/dryad.7d88c
Funding: This work was funded by German Research Foundation (DFG) LI 2360/1-1 (to ML) http://www.dfg.de/, with additional funding from the German Research Foundation and University of Bayreuth, funding programme "Open Access Publishing" http://www.dfg.de/, http://www.uni-bayreuth.de/. The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.
Competing interests: The authors have declared that no competing interests exist.
Tropical forests play a key role in the global carbon cycle storing a total of 471 Pg carbon [1,2]. The soils’ carbon storage potential is generally even greater than that of the vegetation . Don et al.  report, that 36 to 60% of the tropical ecosystem’s carbon is stored in soil. But, land use change from primary forest to other land uses leads to a decrease in soil organic carbon (SOC) stocks [4,5]. Ecuador in particular has the highest annual deforestation rate in South America . Tapia-Armijos et al.  report a reduction of the area covered by natural vegetation by 46% (Southern Ecuadorian provinces). Local farmers make “extensive use of fire” to convert primary forest into farming land and pastures . According to Bahr et al.  9 to 13 Mg SOC per hectare are lost due to land use changes from forest to crop land and pastures. Finally, spatial estimates of SOC are increasingly important to acknowledge the soils’ carbon storage potential in the context of climate change. However, it is particluarly the tropical mountain areas with their thick organic layers which are highly complex and difficult to access . SOC stock data of tropical mountain forest soils are scarce, SOC stock data of the organic layer hardly exist.
Regression based digital soil mapping (DSM) provides a means of regionalising soil data from a limited amount of samples to a landscape level by making use of the factors of soil formation  as predictors. Spatial continuous predictors representing topography and vegetation are obtained from digital elevation models (DEMs) and satellite images. However, for many soil properties, spatial regression modelling may not produce a robust model. According to Ryan et al. , low r2 values may result from one or more of the following causes: (1) poor relation to the available environmental predictor variables, (2) extreme local variation due to unknown or random effects, or (3) the collected data spans a very small interval in the total range of the response variable. While the latter can be mostly avoided by a good sampling design which follows a good representation of the predictor space , the former two causes provide real challenges.
DEMs are often used at their original raster resolution with a 3x3 window size for the calculation of the derived predictors. However, a number of studies suggest that predictor-response relationships are strongly landscape and scale dependent . Cavazzi et al. investigated the interacting effect between window and raster cell size and found cell size to be significant in all considered areas whereas the interaction between window and cell size was significant in morphological rough areas. Finally, soil-forming factors (predictors) vary and respond at different scales . Maynard and Johnson  found a strong scale-dependency for total carbon having the best model performance at coarse neighbourhood extents (150 to 300 m); DEM resolution affected soil-terrain correlations to a much lesser degree. Samuel-Rosa et al.  have shown that investigating the impact of scale in predictors is more important when the predictor-response relationships are weak. Finally, multi-scale as well as feature selection approaches according to  deserve more research to obtain a better prediction accuracy.
The term feature selection refers to the process of removing irrelevant predictors from the predictor set to enhance a model’s performance and generalisation capability (e.g. ). It requires an exhaustive search of all possible subsets of predictors in order to decide which subset performs best. With a large set of predictors, this procedure is simply not applicable . Therefore, due to practicability reasons the selection of a subset which is just good enough if not optimal, might have to be sufficient . Predictor selection procedures can be described by two main categories : (1) filter methods and (2) wrapper methods. Filter methods make a predictor assessment based on general characteristics of the dataset (e.g. predictor–response correlation), independently from the particular machine learning algorithm and hence ignore the predictors’s effectiveness within the particular model. Wrapper methods, evaluate predictor performance by running the particular machine learning algorithm on the dataset . Accordingly, most filter methods evaluate each predictor individually and ignore possibly important predictor interactions, whereas wrapper methods result in an increase in computation time . It is often argued that particularly recursive partitioning methods do not require predictor selection as they are at least theoretically resistant to irrelevant predictors [23,24]. However, Witten and Frank  show that decision trees are affected by non-informative predictors and Kuhn and Johnson  show that ANN and SVM are affected to an even larger extent.
A number of machine learning algorithms are commonly used in DSM, such as e.g. tree-based methods, artificial neural networks (ANN), multivariate adaptive regression splines (MARS) and support vector machines (SVM). Each of them has its strengths and pitfalls and in dependence of the particular application and soil-landscape there is no single algorithm which serves all. Grimm et al.  and Guo et al.  use random forest to map SOC. Stepwise regression was applied by Gessler et al. , Gasparini et al.  and Zhang et al. . Martin et al. [30,31] applied boosted regression trees. Pastick et al.  and Bou Kheir et al.  used decision tree models. ANN were applied by Dai et al. . However, comparisons of various machine learning approaches to spatially predict SOC in tropical mountain areas are scarce. Indeed we only found one: Were et al.  compare SVM to ANN and random forest to map SOC stocks in an afromontane landscape with SVM showing the best performance.
Typically the machine learning algorithms have one or several tuning parameters, and the estimation of these tuning parameters should be based on an estimate of the prediction error . The number of layers and neurons is a crucial decision to be made when constructing ANNs. A network with too few neurons cannot differentiate between complex data patterns, whereas too many neurons would lead to overfitting . In support vector regression the model parameters C and ε as well as the kernel parameters must be tuned to obtain sensible results [37,38]. According to Hastie et al.  random forests do remarkably well with little tuning. In our experience [10,39] tuning random forest and boosted regression tree models can improve prediction results, but tuning needs to be tested carefully for it can also cause overfitting.
Particularly in the complex soil-landscape settings of tropical mountain areas, the choice of methodology for spatial prediction is of high importance to improve the expected poor model results. Accordingly within this study four aspects will be considered to obtain the best possible model:
- Considering different spatial settings/scales for the predictors
- Apply predictor selection strategies to reduce noise and enhance the model’s performance
- Try different machine learning algorithms to capture the complex predictor-response relation. Among the vast choice of algorithms, we selected algorithms that follow different adaptation strategies
- Apply model tuning to optimise the model’s performance
2 Material and Methods
2.1 Research area
The research permit for the respective area was granted by the Ecuadorian Ministry of the Environment. The research area is situated on the eastern escarpment of the southern Ecuadorean Andes between the provincial capitals Loja and Zamora (Fig 1). It forms part of the Natural Reserve Podocarpus—El Condor and is mainly covered by tropical montane forest vegetation changing into shrub and grassland vegetation above the tree line of ca. 2800 m a.s.l.. The area is influenced by an altitudinal decrease in temperature and increase in rainfall. This corresponds to a mean annual air temperature ranging from 19.4°C to 9.4°C and a mean annual total rainfall from 2050 mm to 4400 mm [40,41]. Geologically the area is part of a metamorphic belt of palaeozoic age. Litherland et al.  describe the underlying bedrock as part of the Chiguinda Unit of the litho-tectonic section Loja Terrane, consisting of pelite, schist, phyllite, meta-siltstone, sandstone and quartzite. The area is known for its immense organic layers leading to its description as sloping mire soil-landscape . Its formation is partly related to the dominance of soil hydromorphic properties in the topsoil caused by the silty soil texture, heavy rainfall and shallow slope parallel subsurface flow [43,44]. The occurrence of frequent landslides adds to the complexity of this remote tropical mountain landscape.
2.2.1 Carbon data–sampling and analysis.
The organic layer of the research area was sampled along transects as indicated in Fig 1. The transects were laid to cover the complete hillslope from the ridge to the valley bottom. They were positioned to cover terrain units formed by an overlay of two altitudinal, three slope and two aspect classes, while inclination had to be considered to permit accessibility of the very steep terrain . Each transect was then sampled randomly at three positions to cover the upper, middle and foot slope. Sampling was conducted by using a 20 by 20 cm metal frame of 10 cm height. Samples were oven-dried at 45°C until mass consistency. After the removal of fresh roots, organic carbon contents were determined by a VARIO MAX elemental analyser at the soil laboratory of the University of Halle.
Fig 2 gives an overview of the SOC stocks of the organic layer  used for this study. Interestingly, the highest stocks are found at the upper slope position of transects M1 and M19, both positioned at rather low altitudes, while transects M4 to M7 which are positioned at high altitudes have very low carbon stocks.
Each part of the stacked bar plots refers to the carbon stock of a layer of 10 cm with exception of the lowermost layer, here indicated by black colour.
Predictors used for model development include parameters derived from a satellite image and a DEM. The Landsat 8 OLI/TIRS image from 2014, February 18th, was provided as image courtesy of the U.S. Geological Survey (USGS). The provided image bands were transformed to Top of the Atmosphere Reflectance (TOA) and At-Satellite Brightness Temperature (SatTEMP) using the image’s metadata and information provided by the USGS . The normalised difference vegetation index (NDVI), and the normalised difference moisture index (NDMI) were calculated according to Jackson et al. . The perpendicular vegetation index (PVI) and the transformed soil adjusted vegetation index (TSAVI) are using the soil line concept introduced by Richardson and Wiegand . The soil line concept demonstrates the observed linear relationship between Red and NIR reflectance of bare soil. The parameters β1 and β0 in Eqs 1 and 2 refer to the slope and intercept of this relation and were automatically determined according to Fox et al.  as β1 = 0.71 and β0 = 0.11. X in Eq 2 is a constant usually assumed to be 0.08   .Eq 1Eq 2
All predictors calculated from the mentioned landsat image are shown in Table 1. Landsat bands 1 to 11 refer to the predictor ID 1 to 10, the calculated vegetation indices refer to ID 11 to 14.
Geomorphological and hydrological DEM parameters were calculated by R-package RSAGA [53,54]. The DEM was provided by the DFG RU 816 database . Its cell size was adapted to the landsat image’s cell size of 30 m.
The DEM parameters are listed in Table 2. Please be aware of their ID numbers 15 to 236. Predictors 37 to 236 were calculated with different computing window sizes of 3x3 to 41x41 cells corresponding to different degrees of terrain smoothing and search radii of 45 to 615 m.
2.3 Regression models
Five machine learning algorithms were compared in their performance to predict the organic layer carbon stocks:
- Random Forest (RF)
- Artificial neural network (ANN)
- Multivariate adaptive regression splines (MARS)
- Boosted regression trees (BRT)
- Support vector machine (SVM)
2.3.1 Random forest.
RF is a recursive partitioning method which grows a number of regression trees  and averages the results. In a regression tree the data is subsequently partitioned by the predictor variables into preferably homogeneous subsets regarding the response variable. The mean of each data subset is then used as predicted response value. A partition point or so called node is always defined by that predictor and threshold in its range, which achieves the most homogeneous partition into two subsets (tree branches). At each node all predictors with all possible threshold values are tested.
The RF ensemble model’s stability is obtained through varying the trees’ input dataset and the subset of predictors used to subsequently split the data in each tree node. According to Hastie et al. , RF does not require much tuning. The RF model was adapted by R-package “randomForest” , growing 1000 trees (ntree) and the models default parameters for the size of the predictor subset (mtry = p/3) and the minimum amount of data in the terminal nodes (nodesize = 5). The size of the data subset selected to grow each tree was set to the size of the training set, sampling was done with replacement.
2.3.2 Artificial neural network.
ANNs consist of a number of neurons, the processing units of the algorithm, which are organised in layers. The input–the predictor vectors at each sampling point–has to pass these processing units to relate to the output: the response variable at each sampling point. The neurons of the input layer send data via synapses to the neurons of the first internal layer and these pass it via other synapses to the neurons of the next layer. The synapses store parameters called synaptic weights, which guide the learning process . In general, an ANN is defined by (1) the interconnection pattern between the layers of neurons, (2) the learning rule which updates the weights of these connections, (3) the propagation function that converts a neuron’s weighted input to its output, and (4) the activation function, which determines how the state of the neuron at point of time t+1 is calculated based on its state at point of time t and the new input . Finally, the number of layers and neurons is a crucial decision to be made . Generally one hidden layer is enough to approximate continuous functions, whereas two hidden layers might be necessary in the case of discontinuities . A network with too few neurons cannot differentiate between complex data patterns, whereas too many neurons would lead to overfitting . Data normalisation before ANN training, is essential to prevent larger values from dominating smaller values, which otherwise might lead to premature saturation of neurons .
A two layer feed-forward ANN with Bayesian regularisation  was trained with function “brnn” of R-package “brnn” . The optimal number of neurons (1 to 20) was selected by five repetitions of an external tenfold cross-validation.
For the all-predictor model the maximum number of possible neurons was reduced to 13 (instead of 20) to combat computation time. With a number of 238 predictors, 13 neurons resulted in the estimation of 3094 parameters (weights and biases) and tuning the number of neurons with a range of 1 to 13 neurons in five times repeated tenfold cross-validation with parallel computing still resulted in a computation time of c. 30 hours. In this particular case the tuning resulted in best performance with only 1 neuron.
2.3.3 Multivariate adaptive regression splines.
The model algorithm is constructed by a weighted sum of piecewise linear basis functions of the form (x-t)+ and (t-x)+. Each of these so called linear splines is defined by a predictor (x) and a node (t), with the latter separating the first linear part of the function from the second linear part . Starting with the intercept (the mean of the response variable), the algorithm repeatedly adds such a spline or a product of two or more splines. To define a new spline, at each step, the algorithm searches all predictors and all values of each predictor. This process of adding splines continues until a threshold value for the change in residual error or the maximum number of terms is reached. Due to its construction of splines with their internal node, MARS resembles recursive partitioning algorithms. However, in contrast to the latter, MARS produces continuous models . Alltogether, the model algorithm also resembles stepwise forward linear regression, but instead of using the predictors themselves as input, it also permits to make use of already included model terms and products thereof . MARS was implemented by R-package “earth” , which is based on Friedman’s manuscripts "Multivariate Adaptive Regression Splines" and "Fast MARS" [87,89]. No pruning of the forward pass was applied, but the number of added terms, 1 to 50, was tuned by 5 repetitions of an external tenfold cross-validation. The degree of interaction between predictors was set to 2.
2.3.4 Boosted regression trees.
Despite their many benefits, such as intuitive model structure, handling of input data of different measurement level and scale, insusceptibility to irrelevant predictors, insensitiveness to outliers etc., tree models have also disadvantages limiting their performance, particularly in modelling smooth functions, and the dependence of their structure from the input data .
Boosting sequentially applies a learning algorithm to repeatedly modified data versions and thereby produces a sequence of simple models, whose predictions are finally combined to make the overall prediction . In BRT, at first a regression tree is constructed by assigning all training observations the same weight. Then at each iteration step a new regression tree is trained by giving those observations which impaired model performance in the previous step a higher weight . In this way a sequence of models is build. The resulting BRT model is similar to an additive regression model with the subsequently fitted regression trees as individual terms .
BRT was implemented with R-package “gbm” , which is based on Freund and Schapire’s AdaBoost , using a maximum number of 500 iterations. The optimal number of iterations was then selected by five repetitions of a tenfold cross-validation. Shrinkage, a crucial parameter refering to the learning rate was set to 0.001. The data subsampling rate for each of the trees which incorporates a random effect into the BRT algorithm and provides, therefore, the chance to improve BRT model performance , was set to 0.90. From the two parameters determining the size of each of the regression trees, interaction depth was set to 2 and the minimum number of samples in each of the final subsets was set to 5.
2.3.5 Support vector machine.
SVMs were originally developed as an algorithm for classification, which seperates the classes like other machine learning algorithms by a hyperplane. However, in contrast to other algorithms it searches for that hyperplane which leaves the largest possible margin between classes free of data, leading to a better generalisation probability . Each object to be classified is represented by a vector; the distance of those vectors closest to the hyperplane is maximised. Consequently, only the vectors close to the hyperplane are important for defining it, and are, therefore, called support vectors . Points on the correct side and far away from it are ignored in the process of finding the best hyperplane.
In support vector regression (SVR) there is an analogy in that these “low-error-points” are those with small residuals . The margin of a regression line shall contain all points with errors smaller than ε, i.e. no importance is given to these errors as long as they are smaller than ε . SVR searches for a function which fulfils this ε-criterion and is as flat as possible . Points outside the margin are allowed while a penalty weight C is introduced. It determines the trade-off between allowing points outside ε and the flatness of the regression function . This is particularly important to decrease the impact of outliers. Non-linear regression is typically achieved by applying the so-called Kernel trick. Here, the training data is first transformed into a higher-dimensional feature space by applying a non-linear kernel function. Then a linear model is adapted to this new feature space . Finally, the linear regression in the new feature space is equivalent to a non-linear regression in the original predictor space.
Data normalisation before training a SVM is very important to avoid attributes of higher ranges dominating those with smaller ranges and to avoid problems when calculating the kernel function . The support vector regression model parameters C and ε and the kernel parameters must be tuned to obtain sensible results [37,38]. SVR with a gaussian radial basis function kernel was applied by R-package “e1071” . The width parameter γ, the inverse of the radius of influence of the selected support vectors vi, determines the amount of generalisation of the model. According to Cherkassky and Ma  gives good performance. C and the radial kernel parameter γ were adapted by five repetitions of a tenfold cross-validation. For C a range of 0.1 till 1.0 with a step size of 0.1 was chosen, following Mattera and Haykin  and Cherkassky and Ma . For γ a range of 0.2 till 0.5 with a step size of 0.05 was chosen .
2.4 Model validation and comparison
Five- or tenfold cross-validation are usually recommended to test model performance (e.g. [100,101]). The decision on how the subsets are formed remains important as the formed subsets may have an impact on the cross-validated error . Subsets were formed by random selection. However, to account for the impact of subset membership, tenfold cross-validation was repeated five times with different subsets. Parallel computing of the five repetitions was employed with r-packages “doParallel” and “foreach” [102,103] to combat computation time. Predictors as well as response variable where normalised prior to model training for all five model algorithms to allow for comparison regarding their predictive performance; the scaling default setting, which exists in some of the models (ANN, MARS and SVM), was turned off.
2.5 Predictor selection
Wrapper methods developed for multiple regression models include forward selection, backward elimination, stepwise selection and best subsets. The forward selection procedure usually starts with the predictor most highly correlated with the response variable and tests if the resulting model is significant. It then repeatedly adds predictors at each step, testing for model improvement due to the significance of the calculated F-statistic . We applied a slightly different approach to forward selection: At first all individual predictor models of the five machine learning algorithms were compared in performance by their repeated cross-validated RMSE resulting in a predictor ranking. The single predictor included in the best individual predictor model was then used to start the forward selection procedure. At the end of step 1 it was tested whether the various algorithms’ parameter ranges to select the optimal parameter via cross-validation were reasonable or needed to be adapted.
The model including the ten best predictors of the predictor ranking obtained after step 1 (Fig 3) was compared to two forward selection procedures: Simple forward selection starting with the best predictor (as selected by the best individual predictor model) and a three step forward selection procedure. Fig 3 includes all three procedures:
- Predictor ranking at the end of step 1 (Fig 3) based on all individual predictor models and their repeatedly cross-validated RMSE. Build model of the 10 best predictors as selected by the ten best individual predictor models: 10bestPR.
- Simple forward selection (sFS) includes steps 1 and 2 (Fig 3).
- Three-step forward selection (3stepFS) included all three steps and repeats steps 2 and 3 until no further improvement in repeated cross-validated RMSE is achieved.
3 Results and Discussion
3.1 Selected predictors
Table 3 gives an overview of the selected predictors for each machine learning algorithm (please compare Tables 1 and 2 for predictor IDs). The selected predictors are ordered according to their individual importance (predictor ranking) or order of inclusion during the selection procedures. The best individual predictor is predictor 22 “Standardised Height” in all five model algorithms. As the sFS procedure starts with the best individual predictor, this first predictor is the same as in the model ranking (10bestPR). In two cases this best predictor is removed while applying 3stepFS, the two tree algorithms: RF and BRT. Interestingly, most of the ten single most important predictors (10bestPR) are not included in sFS: only one predictor for RF and MARS, two for ANN and BRT, and three for SVM (indicated by surrounding boxes and bold style in Table 3). For MARS and SVM the 3stepFS procedure did not lead to any improvement compared to sFS, which is why the selected predictors are the same. In the three other model algorithms 3stepFS resulted in an improvement in predictive performance and a reduced number of predictors.
Table 4 includes only those predictors of Tables 1 and 2 which have been selected at least once. To summarise, similar predictors were grouped. This shows that there are many predictors which describe terrain curvature in one way or another, such as Hill Index, Convergence, TS Convexity and the various types of curvature. Filled table cells indicate whether a certain predictor was selected by the predictor selection procedure in the particular machine learning algorithm. The numbers in some of the cells refer to the search radii of the selected predictors. The table shows that sometimes a particular predictor was included in different radii in the forward selection process compared to the 10best predictor ranking. All together convergence and curvature parameters were always among the selected predictors. Another often chosen predictor was “Diffuse Insolation”. The common best predictor for all machine learning algorithms was “Standardised Height”, a predictor referring to the slope position.
Altitude was only important for RF, BRT and SVM, but not for ANN and MARS. Satellite data was important for all algorithms except MARS. The predictor “diffuse insolation” was among the final predictor set (3stepFS selection) in four of the five algorithms. Interestingly, MBI was among the 10bestPR in ANN, MARS and SVM but never selected in any of the two selection procedures in any of the five algorithms. From the hydrological parameters only catchment slope and LS Factor were of any relevance for ANN and MARS.
3.2 Model performance
Fig 4A1–4E1 compare the results of the three selection procedures (1, 2, 3) for the five machine learning algorithms to the all-predictor model (all) results. The figures show the RMSE boxplots of the repeated cross-validation. Please be aware that the RMSEs refer to the normalised C stock data.
RMSE boxplots of repeated cross-validation (a1-e1) and development of the mean RMSE of rpeated cross-validation during the predictor selection process (a2-e2). a) RF, b) ANN, c) MARS, d) BRT, e) SVM. In a1-e1: “all” refers to the all-predictor model, “1”refers to the 10bestPR model, “2” to the sFS model, and “3” to the 3stepFS model. In a2-e2: The star refers to the mean RMSE of the best individual predictor model of step 1, black points refer to added predictors and the resulting mean RMSE during step 2, and grey points refer to the mean RMSE after step 3. The dashed line represents the mean RMSE of the all-predictor model.
The 10bestPR predictor selection improved predictive performance compared to the all-predictor model in all but the ANN algorithm. Applying the sFS procedure improved it even further and in all five algorithms. As mentioned before, in two algorithms, MARS and SVM the subset selection (step 3 in Fig 3) did not improve model results. Therefore, the RMSE boxplots 2 and 3 of Fig 4C1 and 4E1 are exactly the same. But even for ANN and BRT there was only a very slight improvement (Fig 4B1 and 4D1); the median and the interquartile range of the RMSE boxplots are very similar. This improvement is somewhat more pronounced for RF, particularly while comparing the interquartile ranges of RMSE boxplot 2 and 3 (Fig 4A1).
Fig 4A2–4E2 show the development of the repeatedly cross-validated mean RMSE during the three-step selection procedure. Here the dashed line represents the repeatedly cross-validated mean RMSE of the all-predictor model. The star represents the same error measure regarding the best individual predictor model (step 1, Fig 3). The hereby selected first predictor is then entering the simple forward selection procedure (step 2, Fig 3). The first black point refers to the error measure of the model with the first additional predictor (step 2, Fig 3). The number of this first sequence of black points corresponds to the number of predictors added in step 2 before no further improvement can be achieved in sFS. Then the grey point represents the error measure of the best subset selected in step 3 (Fig 3). Black points after the grey point indicate that another sequence of simple forward selection (step 2) is applied and so on. For RF (Fig 4A2) it was enough to run the three steps once, BRT (Fig 4D2) repeated step 2 a second time, ANN needed to repeat step 2 a third time (Fig 4B2). Finally, applying the 3stepFS to BRT resulted in the best predictive performance with a mean RMSE of 0.116, corresponding to a ten-predictor-model (please compare Table 3). However, RF and ANN performed only slightly worse with a mean RMSE of 0.120 and 0.121, respectively. Particularly, the RF model has the benefit of being much less complex with a number of only six predictors (Table 3). Interestingly, the one-predictor model improved model performance compared to the all-predictor model in all but one machine learning algorithm (RF). For MARS this improvement was comparatively high. In all five considered machine learning algorithms sFS resulted in an improvement in the predictive performance. This improvement was particularly pronounced for the three recursive partitioning algorithms: RF, MARS and BRT. Usually, recursive partitioning methods are said to be rather resistant to non-informative predictors . They choose the best predictor at each split and should, therefore, theoretically be resistent to irrelevant predictors. However, adding irrelevant predictors to a classification tree algorithm also resulted in a deterioration by 5 to 10% on datasets tested by Witten and Frank . Witten and Frank  also explain why even relevant attributes can cause harm in recursive partitioning algorithms due to the possibly highly unbalanced subdivision into two subsets. According to Kuhn and Johnson , ANN and SVM are even affected to a much larger extent by irrelevant predictors, which could not be confirmed by our results.
With a high number of non-informative predictors, sFS will always result in an improvement even after few iterations steps. And a number of non-informative predictors is likely to be tested in situations were the predictor-response correlation is weak. During model parameter tuning we further realised that the level of improvement is controlled by the algorithms’ fine tuning. BRT for example resulted in the overall worst model while only 500 iterations were computed, but in the best with the final 10,000 iterations. At the end of step 1 it was tested whether the various algorithms’ parameter ranges to select the optimal parameter via cross-validation were reasonable or needed to be adapted. Accordingly, the upper and lower limit of the SVM gamma and C parameter ranges needed to be adapted with the new range from 10−3 to 104. A higher range of neurons might have resulted into a slightly better RMSE. However, we refrained from doing so since it tremendously increased computation time of the already most time consuming machine learning algorithm. To allow for practicability, the amount of parameter tuning was set to reduce the very time-consuming forward selection (step 2) to a maximum of 12 hours (with parallel computing). This means we do not know if particularly ANN and SVM could be improved even further by testing smaller steps in the C and gamma ranges of SVM and a higher number of neurons in ANN.
3.3 Predicted carbon stocks
Fig 5A shows the median carbon stocks of the soil organic layer under tropical mountain forest and páramo vegetation as predicted by the overall best model: BRT with 3stepFS. The area north of the interprovincial road (Fig 1) was not predicted due to lack of data for this area of high disturbance by fire and non-natural vegetation (pasture). The majority of studies assessing SOC in tropical forests have been conducted in lowland forests, while the SOC stocks of tropical mountain forests are less well known . Moser et al.  investigated SOC in two plots of 20 x 20 m within our research area (at 1890 and 2380 m a.s.l.) and report 3.7 to 4.8 Kg m-2 SOC in the organic layer which lie within the lower range of our findings. The corresponding interquartile range of our predictions (Fig 5B) is below 1 kg m-² for 90% of the area indicating a low prediction uncertainty due to the underlying position specific density function in areas of high SOC stocks. Accordingly, the relative prediction uncertainty in areas of low SOC is higher; rainfall induced landslides, frequently occurring within the area, certainly reduce predictability. The apparently lower organic carbon stocks above an altitude of about 2500 m a.s.l. coincide with a change from mountain forest to shrub and páramo vegetation and the corresponding shallower organic layers (1 to 15 cm) observed during sampling: Transect M4 to M7 (Fig 1). In general, median organic layer thickness varies between c. 20 to 70 cm throughout the area, with organic layer thickness being highest on mid-slope positions .
a) Median prediction value, b) Interquartile range (overlaid hillshading with light source from north).
According to Roman et al. , the high SOC stocks of tropical montane cloud forests cannot only be explained by low soil organic matter turnover due to slow rates of litter decomposition, soil acidity and reduced rates of nutrient cycling. They conclude that the organic soil layer of these ecosystems is highly variable. Furthermore, soil water logging and the altitudinal gradient alone cannot explain the complex spatial pattern . SOC stocks in tropical forest soils of Papua New Guinea are reported to vary between 4.8 and 19.4 kg m-2 (litter and top 100 cm) with an increase of 5.1 kg m-2 per 1000 m increase in altitude, while SOC in litter stayed the same . Don et al.  report SOC stocks between 7.3 and 10.5 kg m-2 in tropical primary forests. They analysed 385 studies on SOC from tropical countries but had to exclude data on organic layer SOC due to dataset scarcity.
The selected predictors which influence the spatial variation include diffuse insolation, topographic ruggedness index, convergence and all considered curvatures (some in different search radii). Diffuse insolation and topographic ruggedness index are related to site exposure, a factor termed important by Roman et al.  when studying the global and local variations in tropical montane cloud forest soils and particularly the accumulation of soil organic matter. The low search radii for plan curvature, cross-sectional curvature, minimum and maximum curvature (75 and 105 m) indicate the importance of local curvatures, while the rather high search radii for the profile and longitudinal curvature (375 to 465 m) indicate the importance of the larger landscape topography. The influence of the diffuse insolation and the influence of the ridge and valley topography on the spatial carbon stock pattern are clearly visible in Fig 5. Carbon stocks are lower in areas of high diffuse insolation (Summit area in the South-East) and in the concave valley structures. Similarly, Wilcke et al. , who studied transects in the lower part of the here investigated area, also report higher humus concentrations on the ridge-tops as compared to the valley structures and explain this with stronger acidification and nutrient leaching on the exposed sites. Gessler et al.  report flow accumulation being a good predictor for SOC. According to them, ignoring the effect of topography presents a serious weakness in approaches to regional and global carbon modelling, as water distribution between convex and concave landscape positions may account for differences in SOC similar to dramatically different climatic zones. According to Grimm et al. , topographical parameters approximate water and solute transport, relate to solar insolation and determine the micro-climate. In that effect, they have the potential to explain large parts of SOC variation.
Predictors obtained from satellite images were not among the selected predictors for the best model: According to Waring and Running  multi-spectral reflectance data obtained from satellite images can provide valuable information on forest structure and productivity. However, in forested areas, remote sensing methods are of limited utility for soil studies due to the variable effects of vegetation .
It was shown, that particularly in the complex soil-landscape setting of tropical mountain areas with possibly low predictor-response correlations, the applied methodology for spatial prediction is of high importance to improve the expected poor model results. Considering different spatial settings/scales for the predictors as well as applying predictor selection and model tuning are important to improve predictive performance.
Concerning predictor selection, choosing predictors individually was not the best option. The forward selection resulted in better predictions than the 10bestPR, emphasizing the importance of predictor interaction. Among the forward selection procedures, the tested 3stepFS algorithm did not result into much improvement compared to the sFS. However, the amount of improvement in predictive performance due to predictor selection depended on model tuning. Finally, even a single- or two-predictor model was better than the all-predictor model independent of the applied machine learning algorithm. Predictor selection is, therefore, also important for recursive partitioning algorithms which are often reportet to be immune to non-informative predictors. To save computation time in the case of many (>100) predictors we recommend to rather apply forward instead of backward selection. Choosing the initial best starting predictor by running the model and not due to correlation seems favourable to make the approach more consistent.
Predictors obtained from satellite images were not among the selected predictors for the best overall model. We assume that satellite images are most probably of higher importance in predicting SOC in areas with contrasting land use types. The highest C stocks appear at the ridges and mid-slope positions at altitudes below 2500 m a.s.l.. Various curvatures in different search radii and the diffuse insolation are best suited to describe this pattern. The interquartile range show slightly bigger uncertainties in the transition zone from montane forest to paramo vegetation at 2500 m a.s.l.. One might argue that other predictors would lead to even better results. However, in most tropical mountain regions it will be difficult to obtain additional predictors apart from the here presented.
This research was funded by the German Research Foundation (DFG) as part of the Platform for Biodiversity and Ecosystem Monitoring and Research in South Ecuador (PAK 825, LI 2360/1-1). Logistic support by the NGO Nature and Culture International (NCI) and the municipal public agency ETAPA is gratefully acknowledged.
Conceived and designed the experiments: ML. Performed the experiments: ML JS. Analyzed the data: ML JS. Contributed reagents/materials/analysis tools: ML JS BG. Wrote the paper: ML JS BG.
- 1. Cerri CEP, Easter M, Paustian K, Killian K, Coleman K, Bernoux M, et al. Predicted soil organic carbon stocks and changes in the Brazilian Amazon between 2000 and 2030. Agric Ecosyst Environ. 2007;122: 58–72.
- 2. Pan Y, Birdsey R, Fang J, Houghton R, Kauppi PE, Kurz W, et al. A large and persistent carbon sink in the world’s forests. Science. 2011;333: 988–993. pmid:21764754
- 3. Post WM, Kwon KC. Soil carbon sequestration and land-use change: processes and potential. Glob Chang Biol. 2000;6: 317–327.
- 4. Don A, Schumacher J, Freibauer A. Impact of tropical land-use change on soil organic carbon stocks—a meta-analysis. Glob Chang Biol. 2011;17: 1658–1670.
Watson RT. Land Use, Land-Use Change, and Forestry [Internet]. 2000. p. 392. Available: http://books.google.com/books?id=EyXQW-gRey0C
FAO. State of the World’s Forests. 2014.
- 7. Tapia-Armijos MF, Homeier J, Espinosa CI, Leuschner C, De La Cruz M. Deforestation and forest fragmentation in south Ecuador since the 1970s - Losing a hotspot of biodiversity. PLoS One. 2015;10: 1–18.
Beck E, Hartig K, Roos K. Forest clearing by slash and burn. In: Beck E.; Bendix J.; Kottke I.; Makeschi F.; Mosandl R, editor. Gradients in a tropical mountain ecosystem of Ecuador Ecological Studies 198. Heidelberg: Springer; 2008.
- 9. Bahr E, Chamba Zaragocin D, Makeschin F. Soil nutrient stock dynamics and land-use management of annuals, perennials and pastures after slash-and-burn in the Southern Ecuadorian Andes. Agric Ecosyst Environ. Elsevier B.V.; 2014;188: 275–288.
- 10. Ließ M, Hitziger M, Huwe B. The sloping mire soil-landscape of southern ecuador: Influence of predictor resolution and model tuning on random forest predictions. Appl Environ Soil Sci. 2014;2014.
Jenny H. Factors of Soil Formation. A System of Quantitative Pedology. New York: Dover Publications, Inc.; 1941.
- 12. Ryan PJ, McKenzie NJ, O’Connell D, Loughhead AN, Leppert PM, Jacquier D, et al. Integrating forest soils information across scales: Spatial prediction of soil properties under Australian forests. For Ecol Manage. 2000;138: 139–157.
- 13. Ließ M. Sampling for regression-based digital soil mapping: Closing the gap between statistical desires and operational applicability. Spat Stat. Elsevier Ltd; 2015;13: 106–122.
- 14. Cavazzi S, Corstanje R, Mayr T, Hannam J, Fealy R. Are fine resolution digital elevation models always the best choice in digital soil mapping? Geoderma. Elsevier B.V.; 2013;195–196: 111–121.
- 15. Behrens T, Zhu AX, Schmidt K, Scholten T. Multi-scale digital terrain analysis and feature selection for digital soil mapping. Geoderma. Elsevier B.V.; 2010;155: 175–185.
- 16. Maynard JJ, Johnson MG. Scale-dependency of LiDAR derived terrain attributes in quantitative soil-landscape modeling: Effects of grid resolution vs. neighborhood extent. Geoderma. Elsevier B.V.; 2014;230–231: 29–40.
- 17. Samuel-Rosa A, Heuvelink GBM, Vasques GM, Anjos LHC. Do more detailed environmental covariates deliver more accurate soil maps? Geoderma. Elsevier B.V.; 2015;243–244: 214–227.
Witten I, Frank E. Data mining–Practical machine learning tools and techniques. 2nd ed. Amsterdam: Elsevier; 2005.
Larose DT. Data mining methods and models. Hoboken: Wiley-Interscience/ John Wiley & Sons, Inc.; 2006.
Xu G, Zong Y, Yang Z. Applied data mining. Booca Raton: CRC Press; 2013.
John G, Kohavi R, Pfleger K. Irrelevant Features and the Subset Selection Problem. Proceedings of the Eleventh International Conference on Machine Learning. 1994. pp. 121–129.
- 22. Kohavi R, John H. Artificial Intelligence Wrappers for feature subset selection. 2011;97: 273–324.
Kuhn M, Johnson K. Applied predictive modeling. New York: Springer Science; 2013. https://doi.org/10.1007/978-1-4614-6849-319
Hastie T, Tibshirani R, Friedman J. The elements of statistical learning. 2nd ed. Springer; 2009. https://doi.org/10.1007/978-0-387-98135-2
- 25. Grimm R, Behrens T, Märker M, Elsenbeer H. Soil organic carbon concentrations and stocks on Barro Colorado Island—Digital soil mapping using Random Forests analysis. Geoderma. 2008;146: 102–113.
- 26. Guo PT, Li MF, Luo W, Tang QF, Liu ZW, Lin ZM. Digital mapping of soil organic matter for rubber plantation at regional scale: An application of random forest plus residuals kriging approach. Geoderma. Elsevier B.V.; 2015;237–238: 49–59.
- 27. Gessler PE, Chadwick OA, Chamron F, Holmes K, Althouse L. Modeling soil-landscape and ecosystem properties using terrain attributes. Soil Sci Soc Am J. 2000; 2046–2056.
- 28. Gasparini P, Di Cosmo L. Forest carbon in Italian forests: Stocks, inherent variability and predictability using NFI data. For Ecol Manage. Elsevier B.V.; 2015;337: 186–195.
- 29. Zhang S, Huang Y, Shen C, Ye H, Du Y. Spatial prediction of soil organic matter using terrain indices and categorical variables as auxiliary information. Geoderma. Elsevier B.V.; 2012;171–172: 35–43.
- 30. Martin MP, Wattenbach M, Smith P, Meersmans J, Jolivet C, Boulonne L, et al. Spatial distribution of soil organic carbon stocks in France. Biogeosciences. 2011;8: 1053–1065.
- 31. Martin MP, Orton TG, Lacarce E, Meersmans J, Saby NPA, Paroissien JB, et al. Evaluation of modelling approaches for predicting the spatial distribution of soil organic carbon stocks at the national scale. Geoderma. Elsevier B.V.; 2014;223–225: 97–107.
- 32. Pastick NJ, Rigge M, Wylie BK, Jorgenson MT, Rose JR, Johnson KD, et al. Distribution and landscape controls of organic layer thickness and carbon within the Alaskan Yukon River Basin. Geoderma. Elsevier B.V.; 2014;230–231: 79–94.
- 33. Bou Kheir R, Bøcher PK, Greve MB, Greve MH. The application of GIS based decision-tree models for generating the spatial distribution of hydromorphic organic landscapes in relation to digital terrain data. Hydrol Earth Syst Sci. 2010;14: 847–857.
- 34. Dai F, Zhou Q, Lv Z, Wang X, Liu G. Spatial prediction of soil organic matter content integrating artificial neural network and ordinary kriging in Tibetan Plateau. Ecol Indic. Elsevier Ltd; 2014;45: 184–194.
- 35. Were K, Bui DT, Dick ØB, Singh BR. A comparative assessment of support vector regression, artificial neural networks, and random forests for predicting and mapping soil organic carbon stocks across an Afromontane landscape. Ecol Indic. Elsevier Ltd; 2015;52: 394–403.
- 36. Basheer IA, Hajmeer M. Artificial neural networks: fundamentals, computing, design, and application. 2000;43: 3–31. pmid:11084225
- 37. Hsu C, Chang C, Lin C. A Practical Guide to Support Vector Classification. BJU Int. 2008;101: 1396–400.
- 38. Cherkassky V, Ma Y. Practical selection of SVM parameters and noise estimation for SVM regression. Neural Networks. 2004;17: 113–126. pmid:14690712
- 39. Hitziger M, Ließ M. Comparison of three supervised learning methods for digital soil mapping: Application to a complex terrain in the Ecuadorian Andes. Appl Environ Soil Sci. Hindawi Publishing Corporation; 2014;2014.
- 40. Fries A, Rollenbeck R, Göttlicher D, Nauss T, Homeier J, Peters T, et al. Thermal structure of a megadiverse Andean mountain ecosystem in southern Ecuador. Erdkd—Arch Sci Geogr. 2009;63: 321–335.
- 41. Rollenbeck R. Variability of precipitation in the Reserva Biólogica San Francisco / Southern Ecuador. Lyonia. 2006;9: 43–51.
- 42. Litherland M, Aspen JA Jamielita R. The metamorphic belts of Ecuador. Overseas Mem Br Geol Surv. 1994;11: 1–147.
- 43. Ließ M, Glaser B, Huwe B. Uncertainty in the spatial prediction of soil texture. Comparison of regression tree and Random Forest models. Geoderma. 2012;170: 70–79.
- 44. Bauer FC. Water flow paths in soils of an undisturbed and landslide affected mature montane rainforest in South Ecuador. 2010; 150.
- 45. Ließ M, Glaser B, Huwe B. Digital soil mapping in Southern Ecuador. Erdkunde. 2009;63: 309–319.
Schmidt J. Statistische Modellierung der Kohlenstoff- und Stickstoffvorräte der organischen Auflage unter Bergregenwald in Südecuador. Universität Leipzig. 2012.
USGS. Using the USGS Landsat 8 product [Internet]. 2013. Available: http://landsat.usgs.gov/Landsat8_Using_Product.php.
- 48. Jackson TJ, Chen D, Cosh M, Li F, Anderson M, Walthall C, et al. Vegetation water content mapping using Landsat data derived normalized difference water index for corn and soybeans. Remote Sens Environ. 2004;92: 475–482.
- 49. Richardson A, Wiegand C. Distinguishing vegetation from soil background information. Photogramm Eng Remote Sens. 1977;43: 1541–1552.
- 50. Fox G, Sabbagh G, Searcy S, Yang C. An Automated Soil Line Identification Routine for Remotely Sensed Images. Soil Sci Soc Am J. 2004;68: 1326.
- 51. Baret F, Guyot G. Potentials and limits of vegetation indices for LAI and APAR assessment. Remote Sens Environ. 1991;35: 161–173.
- 52. Baret F, Guyot G, Major DJ. TSAVI: A Vegetation Index Which Minimizes Soil Brightness Effects On LAI And APAR Estimation. 12th Can Symp Remote Sens Geosci Remote Sens Symp. 1989;3: 1355–1358.
Brenning A. RSAGA: SAGA geoprocessing and terrain analysis in R [Internet]. 2015. Available: http://cran.r-project.org/web/packages/RSAGA/index.html.
Association SUG. SAGA–System for Automated Geoscientific Analysis [Internet]. 2014. Available: http://www.saga-gis.org.
- 55. Nauss T, Göttlicher D, Dobbermann M, Bendix J. Central data services in multidisciplinary environmental research projects. e-Zeitschrift für Agrarinformatik. e-Zeitschrift für Agrar. 2007;2. Available: http://www.preagro.de/ezai/index.php/eZAI/article/view/28/28
- 56. Lyle WZ, Colin RT. Quantitative analysis of land surface topography. Earth Surf Process Landforms. 1987;12: 47–56.
Friedrich K. Digitale Reliefgliederungsverfahren zur Ableitung bodenkundlich relevanter Flächeneinheiten. Frankfurt/M: Frankfurter Geowissenschaftliche Arbeiten D21; 1996.
Friedrich K. Multivariate distance methods of geomorphographic relief classification. In: Heinicke H, Eckelmannn W, Thomasson A, Jones J, Montarella L BB, editor. Land information systems–Developments for planning the sustainable use of land resources European Soil Bureau–Research Report 4. EUR 17729. Office for official publications of the European Communities, Ispra; 1998. pp. 259–266.
- 59. Möller M, Volk M, Friedrich K, Lymburner L. Placing soil-genesis and transport processes into a landscape context: A multiscale terrain-analysis approach. J Plant Nutr Soil Sci. 2008;171: 419–430.
- 60. Rodriguez F. The Black Top Hat function applied to a DEM: A tool to estimate recent incision in a mountainous watershed (Estibère Watershed, Central Pyrenees). Geophys Res Lett. 2002;29: 2–5.
Böhner J, Antonic O. Land surface parameters specific to topo-climatology. In: Hengl T.; Reuter HI, editor. Geomorphometry–Concepts, Software, Applications Developments in Soil Science, Vol 33. Amsterdam: Elsevier; 2009. pp. 195–226.
Oke TR. Boundary layer climates. London: Taylor & Francis; 1987.
Wilson JP, Gallant JC, editors. Terrain analysis–Principles and Applications. New York: John Wiley & Sons, Inc.; 2000.
- 64. Häntzschel J, Goldberg V Bernhofer C. GIS-based regionalisation of radiation, temperature and coupling measures in complex terrain for low mountain ranges. Meteorol Appl. 2005;12: 33–42.
- 65. Anders N, Seijmonsbergen A, Bouten W. Multi-Scale and Object-Oriented Image Analysis of High-Res LiDAR Data for Geomorphological Mapping in Alpine Mountains. Geomophometry 2009. 2009; 61–65.
- 66. Prima ODA, Echigo A, Yokoyama R, Yoshida T. Supervised landform classification of Northeast Honshu from DEM-derived thematic maps. Geomorphology. 2006;78: 373–386.
- 67. Yokoyama R, Shlrasawa M, Pike RJ. Visualizing Topography by Openness: A New Application of Image Processing to Digital Elevation Models. Photogramm Eng Remote Sens. 2002;68: 257–265.
Lea NL. An aspect driven kinematic routing algorithm. In: Parsons A.J.; Abrahams AD, editor. Overland flow: hydraulics and erosion mechanics. London: CRC Press; 1992.
Böhner J, Köthe R, Conrad O, Gross J, Ringeler A, Selige T. Soil regionalisation by means of terrain analysis and process parametrisation. In: Micheli E, Nachtergaele F ML, editor. Soil Classification 2001 European Soil Bureau, Research Report No 7, EUR 20398 EN. 2002.
- 70. Beven KJ, Kirkby MJ. Physically based, variable contibution area model of basin hydrology [Internet]. Hydrological Sciences Bulletin. 1979. pp. 43–69.
- 71. Böhner J, Selige T. Spatial prediction of soil attributes using terrain analysis and climate regionalisation. SAGA—Anal Model Appl Göttinger Aeographische Abhandlungen. 2006;115: 13–28.
- 72. Moore ID, Grayson RB, Ladson AR. Digital Terrain Modelling: a Review of Hydrological, Geomorphological, and Biological Applications. Rev Lit Arts Am. 1991;5: 3–30.
- 73. Desmet P, Govers G. A GIS procedure for automatically calculating the USLE LS factor on topographically complex landscape units. J Soil Water Conserv. 1996;51: 427–433.
- 74. Kinnell PIA. Alternative Approaches for Determining the Usle-M Slope Length Factor for Grid Cells. Soil Sci Soc Am J. 2005;69: 674.
Köthe R, Lehmeier F. SARA–System zur automatischen Relief-Analyse. User Manual. unpublished; 1996.
- 76. Riley S, De Gloria S, Elliot R. A terrain ruggedness that quantifies topographic heterogeneity. Intermt J Sci. 1999;5: 23–27.
- 77. Iwahashi J, Pike RJ. Automated classifications of topography from DEMs by an unsupervised nested-means algorithm and a three-part geometric signature. Geomorphology. 2007;86: 409–440.
Wood J. The geomorphological chracteristics of digital elevation models. University of Leicester. 1996.
Wood J. Geomorphometry in LandSerf. In: Heng T, Reuter H editors. Geomorphometry: Concepts, Software, Applications Developments in Soils Science, 33. Amsterdam: Elsevier; 2009. pp. 333–349.
Breiman L, Friedman J, Olshen R, Stone C. Classification and Regression Trees. Belmont: Wadsworth International Group; 1984.
- 81. Breiman L, Cutler A. Breiman and Cutler’s random forests for classification and regression. Packag “randomForest.” 2012; 29. Available: https://cran.r-project.org/web/packages/randomForest/randomForest.pdf
Lawrence J. Neuronale Netze–Computersimulation biologischer Intelligenz. München: Systhema Verlag; 1992.
Scherer A. Neuronale Netze–Grundlagen und Anwendungen. Vieweg, editor. Braunschweig; 1997.
Masters T. Practical neural network recipies in C++. California: Academic Press; 1993.
- 85. Foresee FD, Hagan MT. GAUSS-NEWTON APPROXIMATION TO BAYESIAN LEARNING ** School of Electrical and Computer Engineering. Network. 1930; 1930–1935.
Perez Rodriguez P, Gianola D. Package “brnn.” 2015;
- 87. Friedman J. Multivariate Adaptive Regression Splines. Ann Stat. 1991;19: 1–67.
Milborrow S. Package “earth”. Multivariate Adaptive Regression Splines [Internet]. 2015. Available: http://cran.r-project.org/web/packages/earth/index.html.
Friedman JH. Fast MARS. Technical Repoert No. 110 [Internet]. 1993. Available: http://www.milbo.users.sonic.net/earth/Friedman-FastMars.pdf.
- 90. Elith J, Leathwick JR, Hastie T, Elith R. Leathwick J., Leathwick & Hastie A working guide to boosted regression trees—Online Appendices Page 1. J Anim Ecol. 2008;77: 802–13. pmid:18397250
Ridgeway G. Package “gbm”. Generalized Boosted Regression Models. 2015; Available: http://cran.r-project.org/web/packages/gbm/index.html.
- 92. Freund Y, Schapire RE. A Decision-Theoretic Generalization of On-Line Learning and an Application to Boosting. J Comput Syst Sci. 1997;55: 119–139.
- 93. Friedman JH. Stochastic Gradient Boosting. Comput Stat Data Anal. 2002;38: 367–378.
- 94. Bennett K, Campbell C. Support vector machines: Hype or hallelujah? SIGKDD Explor. 2000;2. Available: http://www.acm.org/sigs/sigkdd/explorations/issue2-2/bennett.pdf.
Vapnik VN. The nature of statistical learning theory. New York: Springer; 1995.
- 96. Smola A, Schölkopf B. A tutorial on support vector regression. Stat Comput. 2004;14: 199–222.
Olsen D, Delen D. Advanced data mining techniques. Berlin: Springer; 2008.
Meyer D, Dimitriadou E, Hornik K, Weingessel A, Leisch F, Chang C, et al. Package “e1071”. Misc Functions of the Department of Statistics (e1071) [Internet]. E1071. 2015. Available: http://cran.r-project.org/package=e1071
Mattera D, Haykin S. Support vector machines for dynamic reconstruction of a chaotic System. In: Schölkopf B.; Burges C.J.C.; Smola A.., editor. Advances in kernel methods–Support vector learning. Massachusets Institute of Technology; 1999.
- 100. Breiman L, Spector P. Submodel selection and evaluation in regression. The X-random case. Int Stat Rev. 1992;60: 291–319.
- 101. Kohavi R. A Study of Cross-Validation and Bootstrap for Accuracy Estimation and Model Selection. Int Jt Conf Artif Intell. 1995;14: 1137–1143.
Weston S. Package ‘ doParallel '. Foreach Parallel Adaptor for the “parallel” Package [Internet]. 2015. Available: https://cran.r-project.org/web/packages/doParallel/doParallel.pdf.
Weston S. Package “foreach”. Provides Foreach Looping Construct for R [Internet]. 2015. Available: https://cran.r-project.org/web/packages/foreach/foreach.pdf.
- 104. Moser G, Leuschner C, Hertel D, Graefe S, Soethe N, Iost S. Elevation effects on the carbon budget of tropical mountain forests (S Ecuador): The role of the belowground compartment. Glob Chang Biol. 2011;17: 2211–2226.
Roman L, Scatena F, Bruijnzeel L. Global and local variations in tropical montane cloud forest soils. In: Brujinzeel LA, Scatena FN HL, editor. Tropical Montane Cloud Forests International Hydrology Series. New York: Cambridge University Press; 2010.
- 106. Dielemann W, Venter M, Ramachandra A, Krockenberger A, Bird M. Soil carbon stocks vary predictably with altitude in tropical forests: Implications for soil carbon storage. Geoderma. Elsevier B.V.; 2013;204–205: 59–67.
Wilcke W, Boy J, Goller R, Fleischbein K, Valarezo C, Zech W. Effect of topography on soil fertility and water folw in an Ecuadorian lower montane forest. In: Brujinzeel LA, Scatena FN HL, editor. Tropical Montane Cloud Forests International Hydrology Series. New York: Cambridge University Press; 2010.
Waring R, Running S. Forest Ecosystems: Analysis at Multiple Scales. San Diego: Academic Press; 1998.