-
Abbreviations
- APR
- adult plant resistance
- ANN
- artificial neural networks
- AUC
- area under the receiver operating characteristic (ROC) curve
- BL
- Bayesian LASSO
- BRNN
- Bayesian regularized neural network
- BST
- boosting
- CIMMYT
- International Maize and Wheat Improvement Center
- DTH
- days to heading
- GBS
- genotyping-by-sequencing
- GLS
- gray leaf spot
- GS
- genomic selection
- GY
- grain yield
- LR
- leaf rust
- MAS
- marker-assisted selection
- ML
- machine learning
- MLP
- multi-layer perceptron
- PLR
- parametric linear regression
- PNN
- probabilistic neural network
- PS
- phenotypic selection
- QTLs
- quantitative trait loci
- RBFNN
- radial basis function neural network
- RE
- relative efficiency
- RF
- random forests
- RFC
- random forest classifier
- RFR
- random forest regression
- RKHS
- reproducing kernel Hilbert space
- RR
- ridge regression
- SNP
- single nucleotide polymorphisms
- SR
- stem rust
- SVC-l
- support vector linear classifier
- SVC-g
- support vector Gaussian classifier
- SVM
- support vector machine
- SVR-l
- support vector linear regression
- SVR-g
- support vector Gaussian regression
- YR
- yellow rust
- Genomic-enabled prediction
- Machine learning
- Wheat breeding
- Rust resistance
The development of low-cost genotyping strategies such as single nucleotide polymorphisms (SNP) and genotyping-by-sequencing (GBS) (Elshire et al., 2011; Kumar et al., 2012) has made it possible for genomic selection (GS) to offer new possibilities for improving the efficiency of plant breeding methods and programs (Bassi et al., 2016). Crossa et al. (2010), for example, observed in four wheat datasets that the inclusion of markers significantly outperformed (by 8 to 36%) a pedigree-based model in predictive ability. Based on computer simulation studies, Bernardo and Yu (2007) showed that using the whole set of markers available for genotyping achieved better prediction of breeding values than using subsets of markers found to be significantly associated with quantitative trait loci (QTLs), i.e., marker-assisted selection (MAS). This was empirically confirmed by Heffner et al. (2011), who compared phenotypic selection (PS), MAS, and GS prediction performance for 13 phenotypic traits in 374 winter wheat (Triticum aestivum L.) breeding lines. Using GS, prediction accuracy measured by the correlation between observed and predictive values was 28% higher than with MAS, and 95% as high as with PS. Moreover, the mean correlation across six genomic selection indices was 14% higher than for PS. Arruda et al. (2016) found correlations between observed and predicted values that ranged from 0.4 to 0.9 in GS models for Fusarium head blight resistance in wheat, and lower values (r < 0.3) in MAS models. Rutkoski et al. (2014) obtained r ≅ 0.57 using only GBLUP and r > 0.6 using GBLUP after adding a candidate gene-linked marker (Sr2) as a fixed effect (i.e., a combination of GS and MAS).
Rust diseases are one of main causes of wheat production losses throughout the world. Puccinia graminis (stem rust) and Puccinia striiformis (yellow rust) cause major economic losses and, hence, receive attention in wheat breeding programs (Ellis et al., 2014). Stem rust (SR) resistance is generally categorized into two groups: (i) all stage resistance, and (ii) slow rusting, or quantitative adult plant resistance (APR). In the first case, resistance is conferred by race-specific genes and is related to a hypersensitive response; in the second case, slow rusting resistance is usually conferred by multiple loci and not related to a hypersensitive response (Singh et al., 2011). Furthermore, slow rusting quantitative resistance is considered more durable than resistance conferred by pathogen recognition genes and must be improved over multiple selection cycles using screening nurseries for evaluation (Singh et al., 2011).
Strategies such as MAS and GS are alternatives for developing high-yielding wheat with APR. However, the complexity of these traits makes MAS difficult to implement. For instance, the Sr25 gene provides SR resistance only when the Sr2 gene is present (Singh et al., 2011). Also, the number of molecular markers associated with SR resistance genes is not enough for conducting MAS (Singh et al., 2011). Thus, GS is an important option for accumulating favorable alleles for rust resistance (Rutkoski et al., 2011; Ornella et al., 2012). Nevertheless, using GS for predicting disease resistance has its own peculiarities:
1. Linear GS approaches are usually limited to modeling additive effects; however, machine learning (ML) methods are able to include epistasis. Although there is general consensus about the additive nature of APR to rust (Ornella et al., 2012; Rutkoski et al., 2014), some populations show epistasis (Rouse et al., 2014).
2. Most GS applications assume that phenotypic response is continuous and normally distributed, whereas disease resistance is commonly expressed in ordinal scales (e.g., scales from 1 to 9, from 1 to 5, etc.) (Roelfs et al., 1992). Even if the data are transformed, many of the aforementioned problems remain in the model (Gianola 1980, 1982; Kizilkaya et al., 2014; Montesinos-López et al., 2015a). A distinctive attribute of many supervised learning algorithms is that there is no restriction regarding the distribution of response variables. This characteristic makes them less sensitive to the problems that arise in parametric models when ordinal scores are used to quantify diseases. The disadvantages of dealing with count data are discussed by Montesinos-López et al. (2015a, b), who also present an appealing parametric solution to this difficulty.
3. Clearly most economically important traits are affected by large numbers of genes with small effects (e.g., de los Campos et al., 2013), and durable rust resistance in wheat is determined by effective combinations of minor and major genes (Bansal et al., 2014). In this situation, simulation studies suggest the good performance of methods that use variable selection and differential shrinkage of allelic effects (e.g., Bayes B, de los Campos et al., 2013). Hence, ML should provide more flexible methods for genomic-enabled prediction values for SR resistance of wheat lines (González-Recio et al., 2014).
The main objective of this paper is to discuss several state-of-the-art ML methods applied in GS, particularly for predicting wheat diseases. First, we present a brief introduction to ML. Second, we compare results from well-known linear methods (ridge regression, Bayesian LASSO) with those from ML: random forests (RF), support vector machine (SVM) and radial basis function neural network (RBFNN). Third, we examine classification models instead of the regression models commonly used in GS. There are only a few reports of GS being used on rust (and even fewer on ML); however, there is considerable information related to mapping experiments that provide information about the relationship between QTLs and phenotypes. Finally, in the Appendix we present a brief summary of ML methods discussed in this review.
MATERIALS AND METHODS Machine Learning Methods: An OverviewWith advances in GS, volumes of data have dramatically increased, and new research efforts aimed at integrating and unifying several fields of research, such as computer science, ML, bioinformatics, mathematics, statistics, and genetics, have emerged as new data-driven science. This new field of research focuses on estimating more accurate predictive values of unobserved individuals by using statistical learning or ML methods. For example, artificial neural networks (ANN) are common prediction tools in ML. In GS, when a feedforward ANN with a single hidden layer is applied, each marker of the input vector is connected to all neurons in the hidden layer, and these are connected to the output layer with the predicted phenotypic responses. More recently, developments in ANN and faster computer processing have allowed increasing the number of layers to ANN (deep learning) and improving prediction accuracy to capture higher-order interactions between covariates.
The ML is concerned with developing and applying computer algorithms that improve with data (Gareth et al., 2013). Learning can be classified as either supervised or unsupervised. In supervised learning, the objective is to predict a desired output value (trait) inferred from input data. The prediction task is called classification if outputs are categorical (e.g., red-black-blue, or susceptible-moderate-resistant), and regression if outputs are continuous. In unsupervised learning, the objective is to discover groups and associations among input variables where there is no output variable (Hastie et al., 2009).
Many types of methods are used in supervised learning, such as nearest-neighbors methods, decision trees, naive Bayes, Bayes nets, and rule-based learning (Kotsiantis 2007). Methods that have been applied in GS include SVM, RF, and ANN (Gianola et al., 2011; González-Camacho et al., 2012; Pérez-Rodríguez et al., 2012). The reproducing kernel Hilbert space (RKHS), initially presented as a semi-parametric method (Gianola et al., 2006), is now also included in the ML group (González-Recio et al., 2014).
Many ML methods have been implemented in statistical and data-mining open-source software, e.g., Weka (Hall et al., 2009) and R (R Core Team 2016), which run on most modern operating systems (Windows, macOS, and Linux). Because this kind of software is open-source, users can freely modify source codes to fit their own specific needs (Sonnenburg et al., 2007). ML methods have been developed under different theoretical frameworks using classic and Bayesian statistical approaches; they have helped explain field results and focus on the development and improvement of learning algorithms. One example is the theory of probably approximately correct learning described by Valiant (1984) that facilitates the development of boosting algorithms (BST) (Freund and Schapire 1996), with the statistical learning theory being the backbone of SVM (Cortes and Vapnik 1995). The SVM were evaluated for rust resistance (Ornella et al., 2012) and will be discussed in the next section and in the Appendix. To the best of our knowledge, BST has not been tested for GS in rust; however, some authors such as Ogutu et al. (2011) and González-Recio and Forni (2011) reported an outstanding performance of BST on simulated and real data. Boosting is a method for improving the accuracy of regression and classification models (Hastie et al., 2009). Combining multiple learning algorithms helps to improve predictive performance better than any of the constituent learning algorithms alone; the most popular ensemble-based algorithms are bagging, boosting and AdaBoost (Polikar 2006). Sun et al. (2012) proposed a successful ensemble-based approach to imputation of moderate-density genotypes for GS. The RF is another type of ensemble algorithm where the non-parametric function is the average of regression decision trees or classification (Hastie et al., 2009).
Alternatively, BST combines different predictors with some shrinkage imposed on each iteration, given a training dataset (x1, y1), …, (xn, yn), where xi is a set of input vectors and yi is the response variable. The goal is to obtain an approximation of the function that minimizes the expected value of some specified loss function (e.g., mean squared error or exponential loss) over the joint distribution of all (x, y) values. A common way to obtain is using an expansion of the form: [Image Omitted. See PDF] where hm (x, y), the base learner, is a function relating x to y (commonly used base learners are decision trees or ordinary least-squares regression) and βm is a weight assigned to the learner model. This expansion is commonly obtained using gradient descent techniques, i.e., is sequentially updated, such that FM (x) = FM–1 (x) + βM hM (x, y), where βM hM (x, y) is the learner that tends to the gradient of the loss function at a gradient-boosting point; a more detailed description is found in Friedman (2001). González-Recio et al. (2014) applied a BST algorithm for GS. Another BST development is the AdaBoost classifier: once a new base learner is added, the data are reweighed, and then correctly classified cases lose weight, whereas cases that are misclassified gain weight. Thus, next learners focus more on the instances that previous weak learners misclassified. This may cause AdaBoost not to be robust when dealing with noisy data and the presence of outliers (Freund and Schapire, 1996).
Gradient-boosted methods can deal with interactions among several variables, and select the variables automatically. These methods are also robust in the presence of outliers, missing data, and numerous correlated and irrelevant variables, and they take into account the importance of the variables in exactly the same way as RF (Walters et al., 2012).
We also review what are called single learners, i.e., single functions that cannot be decomposed into smaller units. The RKHS and SVM are examples of single learners with good learning performance. The increase of different modeling approaches in GS has made ML converge with classic statistical methods commonly applied in animal and plant breeding (González-Recio et al., 2014).
Parametric Linear Regression (PLR) Versus ML Methods in Genomic Breeding for Rust ResistanceSince GS was originally introduced by Meuwissen et al. (2001), a large number of parametric linear regression (PLR) models have been developed for prediction. The first models proposed by Meuwissen et al. (2001)— e.g., Bayes A and Bayes B— were followed by a plethora of Bayesian regression models— e.g., Bayes C, the Bayesian LASSO and Bayes R (de los Campos et al., 2013; Gianola 2013).
Based on many experiments performed at CIMMYT and elsewhere (Table 1), we observed that several complex traits (e.g., yield and days to heading) are better represented by nonlinear approaches that are able capture small effect cryptic epistatic effects (González-Camacho et al., 2012; Pérez-Rodríguez et al., 2012; Crossa et al., 2014). Pérez-Rodríguez et al. (2012) evaluated 306 CIMMYT wheat lines genotyped with 1717 DArT markers, and days to heading (DTH) and grain yield (GY) traits were measured in 12 environments. The linear models applied were Bayesian LASSO (BL), Bayesian Ridge Regression, Bayes A and Bayes B and the nonlinear models were RKHS, Bayesian regularized neural network (BRNN) and radial basis function neural network (RBFNN). Results show that three nonlinear models had better prediction accuracy than PLR models and that nonlinear models RKHS and RBFNN were consistently better than the linear models. Furthermore, González-Camacho et al. (2012) evaluated BL, RKHS, and RBFNN on 21 maize datasets and also found a slight but consistent superiority of nonlinear models.
Table 1 Articles related to GS and Machine learning methods. Publications, datasets used, traits, models, and performance criteria.
| Publications and datasets used | Traits† | Model‡ | Performance criteria ¶ |
| Ornella et al. (2012) | PBW343xJuchi | BL, RR, SVR-l, SVR-g | r |
| Wheat datasets | PBW343xPavon76 | ||
| PBW343xMuu | |||
| PBW343xKingbird | |||
| PBW343xK-Nyangumi | |||
| González-Camacho et al. (2012); Ornella et al. (2014) | GLS, FFL, MFL, ASI under SS or WW, GY-SS, GY-WW, Stem rust resistance, yellow rust resistance and GY | BL, RR, RKHS, RFR, SVR-l, SVR-g, RFC, SVC-l, SVC-g | r, κ, and RE |
| 14 maize datasets 300 tropical lines genotyped with 55K SNPs | |||
| 16 wheat datasets; 306 wheat lines genotyped with 1717 SNPs | |||
| González-Camacho et al. (2012); maize datasets | GLS, FFL,MFL, ASI, SS or WW | BL, RHKS, RBFNN | r and PMSE |
| González-Camacho et al. (2016); 16 maize datasets; 17 wheat datasets | GLS, GY-SS, GY-WW, GY-LO, GY-HI, FFL, MFL, ASI with SS or WW, DTH | MLP, PNN | AUC and AUCpr |
Traits: female flowering time (FFL), male flowering time (MFL) and the MFL to FFL interval (ASI) under severe drought stress (SS) or in well-watered (WW) environments; grain yield (GY) under SS, WW, low (LO) and high (HI) yielding conditions; days to heading (DTH), gray leaf spot (GLS) resistance.
Models: Bayesian Lasso (BL), Ridge Regression (RR), Reproducing Kernel Hilbert Spaces (RKHS), Random Forest Regression (RFR) and Support Vector Regression (SVR) with linear (l) and Gaussian kernels (g), Random Forest Classification (RFC) and Support Vector Classification (SVC) with linear (l) and Gaussian (g) kernels, radial basis function neural network (RBFNN), multi-layer perceptron (MLP) and Probabilistic neural network (PNN).
Performance criteria: Pearson's correlation coefficient (r), Cohen's Kappa coefficient (κ), relative efficiency (RE), predictive mean squared error (PMSE), area under the receiver operating characteristic curve (AUC) and area under the recall-precision curve (AUCpr).
The additive and oligogenic nature of rust seems to defeat the potential supremacy of non-parametric models (Ornella et al., 2014, Poland and Rutkoski, 2016). Ornella et al. (2012) compared the performance of ridge regression as implemented in the R package rrBLUP (Endelman, 2011), with BL, support vector linear regression (SVR-l), and support vector Gaussian regression (SVR-g) for predicting SR and yellow rust (YR) resistance in five wheat populations. For SR, they analyzed the performance of the models on 10 wheat datasets, each including 90 to 180 individuals (Ornella et al., 2012). The BL and ridge regression (RR) models had similar prediction performance, with a small superiority over SVR models. The BL produced the best results in seven of the 10 datasets. This result supports reports about the additive nature of SR resistance (Singh et al., 2008). Although APR to stem rust is additive in nature, the measurement is semi-quantitative. Although the scale of measurement ranges from 0 to 100% of disease severity, it is comparable to the one through nine scale used to measure the disease. Meanwhile, rust heritability is usually high and major genes play a crucial role in enhancing resistance in CIMMYT wheat germplasm. Further discussion on this topic can be found in the section on classification models.
Regarding YR resistance, for the nine datasets evaluated, predictions were not as good as those for SR resistance, probably because the heritability of YR resistance is lower than the heritability of SR resistance; for example, for population PBW343 × Kingbird, YR resistance had H2 = 0.45, while for SR resistance H2 = 0.90 (Ornella et al., 2012). The superiority of BL was less evident; SVR-l and SVR-g produced the best results (with statistically significant differences) and BL showed the best correlations in three populations.
Rutkoski et al. (2014) evaluated the performance of multiple linear regression models: GBLUP implemented in the rrBLUP package (Endelman, 2011), BL, and Bayes Cπ for predicting adult plant SR resistance in a set of 365 wheat lines characterized by GBS. They also included eight markers (identified in each training dataset by genome-wide association analysis) as fixed effects in the model. The best results were obtained with GBLUP considering markers linked to Sr2 as fixed effects. In Rutkoski et al. (2014), broad-sense heritability was 0.82 and the most strongly significant Sr2 linked marker explained 27% of the genetic variation. Including QTLs with large effects as fixed factors seems promising for PLR models in GS. Arruda et al. (2016) also obtained the best predictive correlations in GS for traits associated with Fusarium head blight resistance in wheat using rrBLUP + “in-house” QTLs (identified by genome-wide association studies, in the same population) treated as fixed effects.
Using simulation, Bernardo (2014) found that major genes should be fitted as having fixed effects in GS, especially if a few major genes are present and if each gene contributes more than 10% of the genetic variance. Higher heritability values (> 0.5) also strengthen the impact of considering major genes as fixed effects. The ML offers flexibility regarding combinations of genes with major and minor effects. Ornella et al. (2014) evaluated six regression models (BL, RR, RKHS, random forest regression [RFR], SVR-l, and SVR-g) for GS in 12 of the 19 rust datasets and the four yield datasets of wheat presented in Ornella et al. (2012). As in previous reports, RKHS had the best results in the four yield datasets, whereas RFR had the best results in nine of the 12 rust datasets (Table 2). This is an expected result, given that due to the additive nature of the trait (Singh et al., 2008; Rutkoski et al., 2014), one would expect PLR models (i.e., BL or RR) to produce the best results; major and minor gene/QTL with different effects are common in rust resistance. However, RFR has the capability of considering markers with large effects (see the Appendix), which is greater than the capability of linear models to capture additive effects. Within the ML group, RFR produced the best results in nine of the 12 rust datasets, RKHS produced the best results in one dataset and SVR-l was the best in one dataset, whereas within the PLR group, only the BL produced the best results in one dataset (Table 2). The performance of ML and PLR models on these rust data still supports the additive nature of rust statistical architecture. Note that RR and SVR-1 use the same regularization term in the cost function, but the algorithm for optimizing the parameters differs in each case because the objective function is different. In RR, a generalized least squares approach is used and in SVR-1, epsilon sensitive optimization is employed.
Table 2 Average Pearson's correlation (of 50 random partitions) of four regression models (RKHS, BL, RFR and SVR-l) applied to 16 wheat datasets. Bold numbers represent the models with the highest average (extracted from Ornella et al. 2014).
| Dataset | RKHS† | BL | RFR | SVR-l |
| KBIRD-Srm | 0.5 | 0.68 | 0.75 | 0.61 |
| KBIRD-Sro | 0.65 | 0.76 | 0.8 | 0.66 |
| KNYANGUMI-Srm | 0.35 | 0.38 | 0.43 | 0.39 |
| KNYANGUMI-Sro | 0.56 | 0.59 | 0.68 | 0.52 |
| F6PAVON-Srm | 0.5 | 0.6 | 0.67 | 0.46 |
| F6PAVON-Sro | 0.57 | 0.67 | 0.71 | 0.54 |
| JUCHI-Ken | 0.28 | 0.32 | 0.22 | 0.42 |
| KBIRD-Ken | 0.16 | 0.21 | 0.45 | 0.17 |
| KBIRD-Tol | 0.4 | 0.49 | 0.53 | 0.43 |
| KNYANGUMI-Tol | 0.14 | 0.28 | 0.49 | 0.21 |
| F6PAVON-Ken | 0.33 | 0.26 | 0.29 | 0.19 |
| F6PAVON-Tol | 0.51 | 0.63 | 0.56 | 0.54 |
| GY-1 | 0.57 | 0.5 | 0.57 | 0.36 |
| GY-2 | 0.49 | 0.49 | 0.45 | 0.36 |
| GY-3 | 0.41 | 0.36 | 0.4 | 0.23 |
| GY-4 | 0.51 | 0.44 | 0.49 | 0.34 |
The models are RKHS (reproducing kernel Hilbert space); BL (Bayesian LASSO); RFR (random forest regression); SVR-l (support vector regression with linear kernel).
It should be mentioned that the precision of the classifiers depends on the number of individuals in a given class, which defines the extreme values. Small thresholds are associated with higher variabilities and worse results in the classifications. Thus by selecting a different α threshold, the conclusions may change. For example, Ornella et al. (2014) gave special emphasis to the analysis for α = 15%, because it is a percentile commonly used in plant breeding programs; however, this threshold may lead to different results as compared to using smaller values of the threshold (e.g., 10 or 5%). In these cases, the problem would generally become more difficult but, in general, the regression approaches could improve, as the problem would start to resemble regression.
Based on our work and the bibliography, we found meaningful consistency in the performance of the methods and the main characteristics of the trait (architecture): a few major genes (rust) versus traits that are the result of the joint action of a large number of genes, each with small effects (e.g., yield), and epistasis versus additivity. In the following paragraphs, we will explore the relationship between the trait's architecture and the performance of some of the best-known GS methods: on the one hand, BL and RR as representatives of linear models; on the other hand, RKHS and RFR as examples of ML methods.
In Fig. 1A, we plot Pearson's correlations (r) of RKHS and RFR versus the relationship in the 12 rust datasets already discussed. The 30 datasets (14 maize trials and 16 wheat trials) and the R and Java scripts used in this work are in the data repository
Fig. 1. Comparison of average Pearson's correlations (of 50 random partitions) of the RKHS and RFR models vs. the average ratio rBLrRR for: (A) 12 rust datasets (B) 4 wheat yield datasets, (C) 12 rust datasets from (A) + 6 GLS datasets, (D) 4 datasets from (B) + 8 non-disease datasets from maize. RKHS (reproducing kernel Hilbert space), BL (Bayesian LASSO), RFR (random forest regression), RR, ridge regression, GLS (gray leaf spot) (adapted from Ornella et al., 2014).
leads to potentially lower accuracy, especially when some large-effect QTLs are present close to or coincide with the markers (Thavamanikumar et al., 2015). Therefore, if there are no QTLs with large effects, the ratio holds
close to one, whereas if there are large-effect QTLs, would be greater than 1; the larger the effect of the QTLs, the greater the deviation from one should be (Thavamanikumar et al., 2015).
This concept was evaluated by analyzing the range of values on the x axis of Fig. 1. The four yield datasets in Fig. 1B present a very narrow interval—from 0.99 to 1— whereas the range of the rust datasets (1A) varied from approximately 0.92 to more than 1.15. As a control, in Fig. 1D we added results of eight non-disease datasets (yield and flowering) from Ornella et al. (2014). Although the range on the x axis was broader, most ratios ranged between 0.97 and 1.01. We also included (Fig. 1C) the results of the five GLS (gray leaf spot) datasets of maize (also analyzed in Ornella et al., 2014), because this disease seems to be caused by a few major genes (Benson et al., 2015). Including these data points does not modify the range on the x axis [0.92, 1.2] nor the slope of the trend-lines. We included the performances of RKHS and RFR on the y axis with the intention of analyzing the response to the above-mentioned ratios of these two versatile methods that take epistasis into account. Interestingly, both methods showed a striking similarity in the slope of the trend-lines in the four datasets. Moreover, if large-effect QTLs are present (Fig. 1A and 1C), RFR outperforms RKHS. The situation is reversed if the trait is controlled by many small-effect QTLs (Fig. 1B and 1D). Although promising and consistent with the theory, our approach requires further research.
To conclude this section, we should mention that there are only a few other reports about GS for rust resistance. Daetwyler et al. (2014), for example, assessed the accuracy of GS for rust resistance in 206 hexaploid wheat (Triticum aestivum) landraces. Based on a five-fold cross-validation, the predicted mean correlations across years were 0.35, 0.27, and 0.44 for leaf rust (LR), stem rust (SR), and YR, respectively, using GBLUP, and 0.33, 0.38 and 0.30, respectively, using Bayes R (e.g., Gianola 2013). Genomic heritabilities estimated using restricted maximum likelihood were 0.49, 0.29, and 0.52 for LR, SR, and YR, respectively. To the best of our knowledge, no ML research has been done on these types of populations.
RESULTS AND DISCUSSION Classification Models in Genomic Breeding for Rust ResistanceAlthough Pearson's correlation is usually applied for assessing the performance of model prediction in GS, it may not be an adequate performance criterion at the tails of the distribution function, where individuals are often selected. The correlation metric is sensitive to extreme values: thus by predicting well the worst and best rust tolerances individuals, a high correlation score will be obtained. This is also what the models optimize for in the training set, when Gaussian noise models are used. The classification methods, on the other hand, use all modeling resources to separate the classes. Thus from a pure ML perspective, when the scores are classification scores (kappa coefficient and relative efficiency), the classification methods should work better. Ornella et al. (2014) also evaluated the performance of the previously described linear and nonlinear regression models for identifying individuals belonging to the best percentile of the distribution function (e.g., α = 15%), using the kappa coefficient (κ) and relative efficiency (RE).
The kappa coefficient is appropriate when data are unbalanced, for it estimates the proportion of cases that were correctly identified by taking into account coincidences expected from chance alone (Fielding and Bell 1997). κ is computed as:[Image Omitted. See PDF] where Po is the agreement between observed and predicted values, computed by where tn is the number of true negatives; tp is the number of true positives; n is the total of individuals; Pe is the probability of agreement calculated as where fp is the number of false positives; and fn is the number of false negatives (Table 3). If there is complete agreement, then κ = 1 if there is no agreement among the raters other than what would be expected by chance, then κ ≤ 0.
Table 3 Confusion matrix for a binary classifier.
tp: true positives, fp: false positives, fn: false negatives, tn: true negatives, and n is total of individuals.
The metric RE based on the expected genetic gain is computed by:[Image Omitted. See PDF] where αʹ and α represent the classes of extreme individuals selected from the observed ranking and predicted values; are the sizes of each class; yi are the observed values; and and is the mean of the test dataset. The numerator in the RE equation represents the selection differential of individuals selected by GS, whereas the denominator represents the selection differential of individuals selected by traditional breeding (Falconer and Mackay, 1996).
Ornella et al. (2014) reported that r was a good performance measure of genomic-enabled prediction when replacing phenotypic selection for α = 15%. For the wheat and maize datasets evaluated in this paper, the relationships between r and κ and between r and RE were very similar (see Fig. 2 and 3 in Ornella et al., 2014). This close relationship is still present if we focus only on the rust data (i.e., with a peculiar distribution of the response variable). In Fig. 2A, we present a comparison of r vs. RE of the regression models (RKHS, RFR, RR, and BL) evaluated on the 12 rust datasets, whereas in Fig. 2B we use r vs. κ (α = 15%) as an example of the same approach and under the same conditions. From both figures, it seems that the ML and PLR models perform approximately the same when selecting the best individuals for a given level of correlation.
Fig. 2. Comparison of average relative efficiency and kappa coefficients (from 50 random partitions) for selecting the best 15% vs. average Pearson's correlations for 4 models evaluated on 12 rust datasets. BL (Bayesian LASSO), RFR (random forests regression), RKHS (reproducing kernel Hilbert space), RR (ridge regression) (adapted from Ornella et al., 2014).
Fig. 3. Support vector machines. Examples of regression (A) and classification (B). (A) Support vector regression ignores residuals smaller than some constant ε (the ε tube) and assigns a linear loss function to larger errors. (B) Support vector classification: training data are mapped from the input space to the feature space, and a hyperplane is used to do the separation.
Classification models have been widely applied in research and industrial settings (Gareth et al., 2013). Classifiers can capture nonlinear relationships between markers and phenotypes, especially in the neighborhood where breeders select the best lines, i.e., where we set the decision threshold. Ornella et al. (2014) also compared the performance of three classifiers for selecting the best individuals: support vector classifier (SVC-l) and (SVC-g) with linear and Gaussian kernels, respectively, and random forest classifier (RFC).
Tables 3 and 4 describe the performance measures (κ and RE) of the proposed regression and classification models for selecting the best 15% of individuals in the 16 wheat and 12 rust datasets. The SVC-l obtained the best κ in the SR datasets, in five YR datasets and in the GY-1 grain yield dataset (Table 4). This classifier outperformed F6PAVON-Srm and KNYANGUMI-Srm in the SR datasets, and F6PAVON-Ken in the YR and grain yield datasets. The SVC-g yielded the best κ in one YR dataset (KBIRD-Ken). Ridge regression had the best κ in GY-2, and RKHS gave the same value as BL in GY-3 and a higher value in GY-2 (Table 4).
Table 4 Average kappa coefficients (of 50 random partitions) of three regression models (RKHS, BL, RR) and two classifiers (SVC-g, SVC-l) applied to 16 wheat datasets when the best 15% of the individuals are selected. Bold numbers indicate models with the highest average (extracted from Ornella et al., 2014).
| Dataset | RKHS† | BL | RR | SVC-g | SVC-l |
| KBIRD-Srm | 0.10 | 0.30 | 0.28 | 0.24 | 0.44 |
| KBIRD-Sro | 0.32 | 0.39 | 0.34 | 0.17 | 0.53 |
| KNYANGUMI-Srm | 0.17 | 0.18 | 0.21 | 0.23 | 0.42 |
| KNYANGUMI-Sro | 0.44 | 0.49 | 0.41 | 0.33 | 0.49 |
| F6PAVON-Srm | 0.26 | 0.26 | 0.25 | 0.36 | 0.46 |
| F6PAVON-Sro | 0.29 | 0.38 | 0.35 | 0.26 | 0.46 |
| JUCHI-Ken | 0.26 | 0.26 | 0.26 | 0.24 | 0.28 |
| KBIRD-Ken | 0.03 | 0.08 | 0.03 | 0.23 | 0.21 |
| KBIRD-tol | 0.37 | 0.35 | 0.33 | 0.30 | 0.41 |
| KNYANGUMI-tol | 0.03 | 0.06 | 0.06 | 0.29 | 0.33 |
| F6PAVON-Ken | 0.16 | 0.16 | 0.11 | 0.29 | 0.35 |
| F6PAVON-tol | 0.38 | 0.32 | 0.31 | 0.39 | 0.42 |
| GY-1 | 0.23 | 0.14 | 0.14 | 0.27 | 0.40 |
| GY-2 | 0.25 | 0.24 | 0.26 | 0.24 | 0.18 |
| GY-3 | 0.23 | 0.23 | 0.23 | 0.22 | 0.15 |
| GY-4 | 0.42 | 0.35 | 0.34 | 0.35 | 0.30 |
The models are RKHS (reproducing kernel Hilbert space), BL (Bayesian LASSO), RR (ridge regression), SVC (support vector classification) with Gaussian (g) or linear (l) kernels.
Regarding RE (Table 5), SVC-l gave the highest RE in four SR datasets (both KBIRDs, KNYANGUMI-Srm and F6PAVON-Srm) and in all YR datasets except one (KBIRD-tol) (the best RE was RFR). The BL had the highest RE values in two SR datasets (KNYANGUMI-Sro and F6PAVON-Sro), whereas RKHS gave the best RE values in GY-2, GY-3, and GY-4. Finally, SVR-g had the best average RE in GY-1.
Table 5 Average relative efficiency (of 50 random partitions) of three regression models and two classifiers (RKHS, BL, RFR, SVC-g, and SVC-l) applied to 16 wheat datasets when the best 15% of the individuals are selected. Bold numbers represent the highest values (extracted from Ornella et al., 2014).
| Dataset | RKHS† | BL | RFR | SVC-g | SVC-l |
| KBIRD-Srm | 0.28 | 0.60 | 0.56 | 0.12 | 0.71 |
| KBIRD-Sro | 0.62 | 0.76 | 0.81 | 0.47 | 0.82 |
| KNYANGUMI-Srm | 0.51 | 0.60 | 0.67 | 0.50 | 0.81 |
| KNYANGUMI-Sro | 0.65 | 0.74 | 0.68 | 0.48 | 0.73 |
| F6PAVON-Srm | 0.58 | 0.61 | 0.64 | 0.57 | 0.78 |
| F6PAVON-Sro | 0.61 | 0.75 | 0.69 | 0.49 | 0.74 |
| JUCHI-Ken | 0.50 | 0.50 | 0.17 | 0.26 | 0.55 |
| KBIRD-Ken | 0.08 | 0.27 | 0.39 | 0.19 | 0.48 |
| KBIRD-tol | 0.46 | 0.52 | 0.64 | 0.44 | 0.62 |
| KNYANGUMI-tol | 0.09 | 0.20 | 0.32 | 0.34 | 0.55 |
| F6PAVON-Ken | 0.32 | 0.21 | 0.27 | 0.26 | 0.57 |
| F6PAVON-tol | 0.58 | 0.57 | 0.55 | 0.48 | 0.64 |
| GY-1 | 0.53 | 0.38 | 0.52 | 0.62 | 0.48 |
| GY-2 | 0.50 | 0.46 | 0.39 | 0.45 | 0.32 |
| GY-3 | 0.46 | 0.45 | 0.37 | 0.23 | 0.34 |
| GY-4 | 0.59 | 0.48 | 0.56 | 0.47 | 0.36 |
The models are RKHS (reproducing kernel Hilbert space), BL (Bayesian LASSO), RFR (random forest regression), SVC (support vector classification) with Gaussian (g) or linear (l) kernels.
The binary classification model was better than the regression models on disease datasets probably due to the skewness of the distribution (Ornella et al., 2014; Montesinos-López et al., 2015a) or to how regression and classification models accounted for the complexity of the relationship between genotypes and phenotypes.
Concerning the impact of the distribution of the response variable (i.e., a discrete instead of a continuous variable) on the accuracy of prediction, Montesinos-López et al. (2015a, b) give several examples. In both papers, the authors introduced methods for dealing with discrete instead of continuous variables. Since these methods can be considered part of the PLR group, we refer the interested reader to the original references. For simplicity, we will finish this presentation by describing the ability of ANN to perform accurate multiclass classification.
González-Camacho et al. (2016) extended the work of Ornella et al. (2014) on assessing the performance of multi-layer perceptron (MLP) and probabilistic neural network (PNN) classifiers for estimating the probability of one individual belonging to a desired category. These classifiers were assessed using continuous traits grouped into three classes based on the empirical distribution function of traits. This is relevant for scales such as susceptible-moderate-resistant and can also be extended to more classes. González-Camacho et al. (2016) used genomic and phenotypic data from 16 maize datasets and 17 wheat datasets in various trait-environment conditions. The best individuals were selected from 15 and 30% of the upper class and similarly for the lower class. Wheat datasets were also evaluated as a binary class for the same percentiles in the upper and lower classes.
The performance of the two classifiers was assessed with two metrics: the area under the receiver operating characteristic (ROC) curve (AUC) and the area under the precision-recall curve (AUCpr) (Murphy 2012). The AUC produces a value between 0 and 1. The ROC curve is the plot of recall (R) versus the false positive rate (fpr) for a set of thresholds τ given.
Recall is the fraction of individuals correctly classified with respect to the observed individuals and is computed as:[Image Omitted. See PDF]
The AUCpr measure is more informative with unbalanced classes (Keilwagen et al., 2014). The AUCpr is the area under the precision-recall curve, i.e., the plot of precision P vs. R for a set of thresholds τ. Precision P is computed as[Image Omitted. See PDF]
The PNN outperformed the MLP classifier for selecting on 15 and 30% upper classes, in 7 environments for GY, and for selecting on 15 and 30% lower classes in 10 environments for DTH (González-Camacho et al., 2016). In GS, it is usual that p markers >> n phenotype individuals, and PNN is promising because it has better generalization ability than MLP, with the advantage of being computationally faster than MLP in achieving optimal solutions. We must recall that, despite their good performance, both methods consider categorical response variables. Still to be explored are several methods in the ML repository that also take advantage of ordinal response variables (Hall and Frank, 2001).
CONCLUSIONSThe features of the architecture of rust resistance— i.e., their additive nature, the effective involvement of major and minor genes and the nature of the response variable (finite and discrete)— demand specific methods for dealing with the aforementioned characteristics. The flexibility of ML methods suggests ML is a valuable alternative to well-known parametric methods for predicting categorical and continuous responses in genomic selection.
We compared the performance of several regression/classification models against some parametric models (BL, ridge regression, etc.) on SR and YR. To show the broad horizon of ML capabilities, we also discussed the performance of ML methods on other traits, e.g., yield and days to heading in wheat or gray leaf spot resistance in maize. Results confirmed that ML methods could circumvent restrictions imposed by the statistical architecture of the trait. Further development is needed to adapt new approaches to wheat populations and/or environments.
Conflict of Interest DisclosureThe authors declare that there is no conflict of interest.
The RKHS was initially proposed for animal breeding by Gianola et al. (2006); de los Campos et al. (2009) presented a genetic evaluation method where the values can be used for genetic evaluation using dense molecular markers and pedigree. Furthermore, de los Campos et al. (2009) showed that the statistical models routinely used in genomic evaluations are particular cases of RKHS. De los Campos et al. (2010) give an overview of RKHS methods, as well as some ideas on how to set the parameters of the models and how to select the reproducing kernel; they also present an application in plant breeding. Other authors (e.g., Crossa et al., 2010; González-Camacho et al., 2012) have successfully applied RKHS models in plant breeding. The RKHS is the core mathematics used by all kernel methods; it is also the basis for the SVM and SVR methods, among others. Here the RKHS method additionally uses multiple kernel learning, which is an extension that combines several kernel functions.
A kernel is any smooth function K that defines a relationship between pairs of individuals through a covariance or by using similarity functions that can be defined in many ways. Several RKHS models have been proposed for GS (González-Recio et al., 2014). Here we describe the RKHS strategy with “kernel averaging” that was introduced in the field of quantitative genetics by de los Campos et al. (2010). The regression function takes the following form:[Image Omitted. See PDF] where are vectors of dimension p markers; αi are regression coefficients; and K (.) is a positive definite function (the reproducing kernel, RK) evaluated in a pair of lines denoted by i and i′. In the case of a Gaussian kernel, , where is the squared Euclidean distance between the vectors of markers for a pair of individuals (i, i’); and h is a parameter that is known as bandwidth, h > 0. The RK provides a set of n basis functions, which are nonlinear on marker genotypes; but the regression function is obtained by linearly combining the basis functions that are generated by the RK, with weights given by the corresponding regression coefficients.
To avoid over-fitting, the vector of regression coefficients is estimated using penalized or Bayesian methods. The set of basis functions is defined a priori via the choice of kernel; if the kernel is not selected properly, the ability of RKHS to capture complex patterns is affected. De los Campos et al. (2009) proposed kernel averaging to increase the accuracy of prediction. This multikernel approach was implemented using a Bayesian approach; for example, for three kernels:[Image Omitted. See PDF] where is a standardized squared Euclidean distance, Vj is the sample variance of the j-th marker, q05 is the fifth percentile of , and h1 = 5; h2 = 1; h3 = are values of h, such that K1 (…) gives local basis functions and K3(…) gives basis functions with a wider span.
A1. Support Vector Regression (SVR)Support vector machine learning can be applied to classification or to regression problems (Cortes and Vapnik, 1995). Support vector regression (SVR) is a supervised learning algorithm capable of solving complex problems. The SVR aims to learn an unknown function based on the structural risk minimization principle. The SVR considers, in general, approximating functions of the form . When we consider a linear regression model:[Image Omitted. See PDF]
In regression, usually an error of approximation is used instead of the margin between an optimal separating hyperplane and support vectors like in SVM. A linear loss function with ε-insensitivity zone is given by (Cortes and Vapnik 1995):[Image Omitted. See PDF]
When the difference between the measured and predicted values is less than ε, then the error is equal to zero. The ε-insensitive loss function defines an ε tube as depicted in Fig. 3. When the predicted response is within the tube, the error is zero. For points outside the tube, the error is the difference between the predicted response and the radius ε of the tube.
In an SVR problem, the objective is to minimize the empirical risk and the squared norm of the weights simultaneously. That is, to estimate the hyperplane f(x,w) = xTw + b by minimizing the empirical risk:[Image Omitted. See PDF]
Minimizing R is equivalent to minimizing the risk as a function of slack variables (Cortes and Vapnik, 1995):[Image Omitted. See PDF] subject to constraints,[Image Omitted. See PDF] where ξ and ξ* are slack variables. ξ = y − f (x,w) − ε for points above an ε tube; ξ*= f (x,w) − y − ε for points below an ε tube; the constant C is a parameter defined by the user. Big C values restrict large errors, reduce the approximation errors and increase the weight vector norm || w ||. The latter increase does not favor good generalization performance of a model; ε is another design parameter that defines the size of an ε tube. Figure 3(A) describes a regression problem for which no linear solution exists, given the radius of an ε tube.
To perform a nonlinear regression, nonlinear functions f are defined by mapping the training dataset into a higher dimensional space, the feature space F (φ: X →F), and then solve the linear regression problem there (Fig. 3B). As in a classification problem, to make this approach computationally feasible, a symmetric kernel function can be used K(x;y) = [ϕ(x)]∙ ϕ(y)] that directly creates a dot product in the feature space (Hastie et al., 2009).
The linear kernel defined as K(x;y) = (x × y) is the simplest kernel method. Another kernel method is the previously presented Gaussian kernel. The best values of C, ε and, eventually, h (Gaussian kernel) are commonly estimated by means of a grid search combined with a cross-validation procedure (Maenhout et al., 2007).
A2. Feedforward Artificial Neural Networks (ANN)The ANN has the ability to capture linear and nonlinear relationships between predictor variables and response variables, including interactions between explanatory variables (Gianola et al., 2011; González-Camacho et al., 2012). The ANN used in GS are based on a single hidden layer for predicting phenotypic responses from genotypic inputs. The first layer contains the marker input vector; the hidden layer contains a varying number of neurons with nonlinear transfer functions. Each neuron in the hidden layer transforms the input and sends the result to the output layer. Thus the predicted responses of the neural network represent the weighted outputs from each neuron.
The ANN have internal parameters with two main model choices: (i) the number of neurons in the hidden layer, and (ii) the type of transfer function applied in each neuron. A low number of neurons in a neural network may not be able to capture complex patterns. On the other hand, a neural network with many neurons can have over-fitting and poor predictive ability (Haykin, 1994).
The multilayer perceptron (MLP) and the radial basis function neural network (RBFNN) are the most frequently used feedforward neural network models. The MLP uses the back-propagation algorithm to estimate the weight vectors. In this case, the input data are iteratively given to the neural network. For each input data, the error between the predicted and desired response is computed. The error is then back-propagated to the ANN and employed to modify the weights; thus the error diminishes at each iteration. The RBFNN is defined by a hidden layer of processing units with Gaussian radial basis functions. Its structure for obtaining a phenotypic response from a marker input vector is described in Fig. 4 (González-Camacho et al., 2012).
Fig. 4. Structure of a radial basis function neural network (RBFNN). In the hidden layer, each input vector (Xi1,…,Xip) is summarized by the Euclidean distance between the input vectors xi and the centers cm, m = 1, …, M neurons, i.e., hm||xi – cm||, where hm is a bandwidth parameter. Then distances are transformed by the Gaussian kernel exp(−(hm‖ xi−cm ‖)2) for obtaining the responses yi=w0+∑m=1Mwmzmi+εi, (extracted from González-Camacho et al. 2012).
An MLP classifier maps a set of input data into C different disjoint classes (Fig. 5). In general, MLP is flexible because no assumptions are made about the joint distribution of inputs and outputs. The hidden layer contains M neurons. In each one, an S score is obtained using a linear combination of the input markers plus an intercept term, i.e.,[Image Omitted. See PDF]
Fig. 5. Structure of a multi-layer perceptron (MLP) classifier. The hidden layer has M neurons whose outputs are computed by a linear combination of the input and weight vectors plus a bias term. In the output layer, the output of each of the S neurons (classes) is computed by a nonlinear transfer function and the phenotypic response is regressed from the data-derived features (extracted from González-Camacho et al. 2016).
Then S is transformed using a nonlinear transfer function (e.g., a hyperbolic tangent sigmoid function), which maps from the real line to the interval [–1,1]. The sum layer contains C neurons (i.e., the number of classes) and each neuron processes the outputs of the hidden layer also using a linear combination of the outputs plus the intercept and a transformation using the tangent sigmoid function. Finally, the output of the MLP is a column vector of C elements. The index of the largest element in the vector specifies which of the C classes the vector represents (González Camacho et al., 2016).
The mean squared error between the predicted class Ĉ and the desired class C is commonly used to optimize an MLP classifier. Ĉ is a matrix of size S × n, with columns containing values in the [0,1] interval. Training of an MLP involves estimating all the parameters using the backpropagation method, based on the conjugate gradient method (Møller, 1993).
A4. Probabilistic Neural Network (PNN) ClassifierA PNN has a single hidden layer (Fig. 6), but conceptually it is different from MLP. The pattern layer has M neurons (M = number of individuals in the training dataset) and calculates the Euclidean norm between the input vector xi and the center vectors cm using a Gaussian kernel. The output of this pattern layer is a vector ui with M elements, where[Image Omitted. See PDF] and h is the spread of the Gaussian function, showing how near xi is to cm (the mth training data). At that point, in the summation-output layer, each umi is transformed into a vector whose elements are defined by . The zmi values are then used to obtain the contribution for each k class, that is, , where wkm are weights computed from the desired class C matrix of dimension S × n to produce a vector of probabilities ĉi = softmax (vi) of size S × 1 as its response; the softmax function σ (.) is given by[Image Omitted. See PDF] where vi is a desired vector of size S × 1. For each k class, the softmax function transforms the outputs of processing units on the output layer in the [0, 1] range (González Camacho et al., 2016).
Fig. 6. Structure of a probabilistic neural network (PNN) classifier with a single hidden or pattern layer. In this layer, Euclidean distances are computed using a Gaussian kernel. Then the summation-output layer integrates the contributions for each class to generate a vector of probabilities (extracted from González-Camacho et al. 2016).
An advantage of the PNN is that, for a given a value of h, it requires a single iteration over all the xi in the training set. Furthermore, PNN produces as output (ci) the posterior probabilities of each class membership within a short training time. The PNN converges to a Bayes classifier if the training dataset is large enough (Wasserman, 1993).
A5. Deep LearningConventional ANN models contain at least one hidden layer. Reaching beyond, deep learning architectures use many layers of nonlinear processing units (Goodfellow et al., 2016). The presence of more than one hidden layer allows the network to learn higher order interactions. The computing of the weight vector in each layer combines simple and complex features so that the most suitable hierarchical representations can be captured from the data.
Min et al. (2016) characterized traditional deep learning architectures into six types: deep neural network (DNN), multilayer perceptron (MLP), stacked auto-encoder, deep belief network, convolutional neural network (CNN) and recurrent neural network (RNN). They also defined three emergent architectures: convolutional auto-encoder, multi-dimensional recurrent neural networks, and deep spatiotemporal neural networks.
Each architecture has advantages. For instance, CNN is suitable for studying spatial information, DNN is suitable for analyzing internal correlations in high-dimensional data and RNN is suitable for analyzing sequential information.
A key element to take into account when training deep learning architectures is regularization, whose objective is to avoid over-fitting and attain good generalization performance. The two most common methods are weight decay, where a penalty term is imposed on the error loss function so that the weight vectors converge to smaller absolute values, and dropout, which randomly removes hidden units from the network.
A6. Random Forests (RF)The RF is a non-parametric approach for solving regression and classification problems proposed by Breiman (2001) based on bagging, i.e., “bootstrap aggregated sampling.” Gianola et al. (2014) applied the concept of bagging to GBLUP. The RF captures complex interactions and it is robust to over-fitting the data. However, variable importance measures are reported to be systematically affected by minor allele frequency (Walters et al., 2012). The RF combines the output of decorrelated decision trees from classification or regression generated from bootstrapped samples of the training dataset (Hastie et al., 2009).
For both classification and regression models, each tree is constructed using the following heuristic algorithm:
i. Draw samples with replacements from the entire dataset by bootstrapping so that records of the ith individual appear several times, or not at all, in the bootstrapped set. Root-node is the name given to each subset of the bootstrapped sample.
ii. Draw a certain number (mtry) of input variables (SNPs) at random, and select the jth SNP j = 1,…,mtry that minimizes a loss function. The entropy criterion is used for classification problems, whereas the mean squared error is used for regression problems.
iii. According to the genotypes of the jth SNP, the data in the node are separated into two new subsets.
iv. Repeat steps ii–iii for each new node with the data until a minimum node size (number of individuals) is reached. This number is usually five or less.
v. Construct a number ntree (> 50 – 100) of new trees by repeating steps I through iv and using new bootstrapped samples.
Finally, to make a prediction using a new example, the RF combines the outputs of the classification or regression trees based on bootstrapped samples of the dataset. In classification, the class of an unobserved example is predicted by counting the number of votes (usually one vote per decision tree is used) and assigning the class with the highest number of votes (Liaw, 2013). In regression, the method averages the ntree outputs.
Two main features can be tuned in RF:
1. Number of covariates (markers) sampled at random for each node. Although cross-validation strategies can be used to optimize mtry, the default values of mtry = p / 3 (p is the number of predictors) when building a regression tree, or mtry = when building a classification tree, are commonly used (Gareth et al., 2013).
2. Number of trees; there is consensus that RF does not overfit with a growing number of trees; however, building each tree is very time-consuming. In our experience, an ntree range between 500 and 1000 is safe enough for plant datasets evaluated in GS.
Another key benefit of RF, as implemented in the R package “randomForests” (Liaw 2013), is that it provides two different measures for judging the relevance of predictor variables (e.g., marker or environmental effects). The first measure is computed by permuting markers from the OOB (“out of bag” sample) data, i.e., those individuals (around a third of the total) that were not selected in the bootstrapped sample. Briefly, after a tree is constructed, the OOB is passed down the tree and the prediction accuracy on this sample is calculated using several criteria (error rate for classification, mean squared error for regression problems). The genotypes for the p-th SNP are permuted in the OOB and the sample is again passed down the tree. The relative importance is computed as the difference between the prediction accuracy of the original OOB and the prediction accuracy of the OOB with the permuted variable. Then this step is repeated for each covariate (SNP) and the decrease in accuracy is averaged over all trees in the random forests (normalized by the standard deviation of the differences). The SNPs that show a larger decrease are assumed to be more relevant (González-Recio et al., 2014).
The second measure is the reduction in impurities of the node from splitting the variable, averaged over all trees. For classification problems, node impurity is measured by the Gini index, whereas for regression, it is measured by the residual sum of squares. Impurity is calculated only at the node at which that variable is used for that split (Liaw, 2013).
You have requested "on-the-fly" machine translation of selected content from our databases. This functionality is provided solely for your convenience and is in no way intended to replace human translation. Show full disclaimer
Neither ProQuest nor its licensors make any representations or warranties with respect to the translations. The translations are automatically generated "AS IS" and "AS AVAILABLE" and are not retained in our systems. PROQUEST AND ITS LICENSORS SPECIFICALLY DISCLAIM ANY AND ALL EXPRESS OR IMPLIED WARRANTIES, INCLUDING WITHOUT LIMITATION, ANY WARRANTIES FOR AVAILABILITY, ACCURACY, TIMELINESS, COMPLETENESS, NON-INFRINGMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Your use of the translations is subject to all use restrictions contained in your Electronic Products License Agreement and by using the translation functionality you agree to forgo any and all claims against ProQuest or its licensors for your use of the translation functionality and any output derived there from. Hide full disclaimer
© 2018. This work is published under http://creativecommons.org/licenses/by-nc-nd/4.0/ (the “License”). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.
Abstract
New methods and algorithms are being developed for predicting untested phenotypes in schemes commonly used in genomic selection (GS). The prediction of disease resistance in GS has its own peculiarities: a) there is consensus about the additive nature of quantitative adult plant resistance (APR) genes, although epistasis has been found in some populations; b) rust resistance requires effective combinations of major and minor genes; and c) disease resistance is commonly measured based on ordinal scales (e.g., scales from 1–5, 1–9, etc.). Machine learning (ML) is a field of computer science that uses algorithms and existing samples to capture characteristics of target patterns. In this paper we discuss several state‐of‐the‐art ML methods that could be applied in GS. Many of them have already been used to predict rust resistance in wheat. Others are very appealing, given their performance for predicting other wheat traits with similar characteristics. We briefly describe the proposed methods in the Appendix.
You have requested "on-the-fly" machine translation of selected content from our databases. This functionality is provided solely for your convenience and is in no way intended to replace human translation. Show full disclaimer
Neither ProQuest nor its licensors make any representations or warranties with respect to the translations. The translations are automatically generated "AS IS" and "AS AVAILABLE" and are not retained in our systems. PROQUEST AND ITS LICENSORS SPECIFICALLY DISCLAIM ANY AND ALL EXPRESS OR IMPLIED WARRANTIES, INCLUDING WITHOUT LIMITATION, ANY WARRANTIES FOR AVAILABILITY, ACCURACY, TIMELINESS, COMPLETENESS, NON-INFRINGMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Your use of the translations is subject to all use restrictions contained in your Electronic Products License Agreement and by using the translation functionality you agree to forgo any and all claims against ProQuest or its licensors for your use of the translation functionality and any output derived there from. Hide full disclaimer
Details
1 Statistics and Computer Science Graduate Program, México, CP
2 NIDERA, Tuerto, Argentina
3 Dep. of Animal Sciences, Dairy Science, and Biostatistics and Medical Informatics, Univ. of Wisconsin, Madison, WI
4 Biometrics and Statistics Unit, International Maize and Wheat Improvement Center (CIMMYT), México City





