Academic Editor:Jennifer Wu
Faculty of Information Technology & State Key Laboratory of Quality Research in Chinese Medicines, Macau University of Science and Technology, Avenida Wai Long, Taipa 999078, Macau
Received 24 October 2014; Revised 5 April 2015; Accepted 30 April 2015; 16 June 2015
This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.
1. Introduction
Identifying molecular biomarker or signaling pathway involved in a phenotype is a particularly important problem in genomic studies. Logistic regression is a powerful discriminating method and has an explicit statistical interpretation which can obtain probabilities of classification regarding the class label information.
A key challenge in identifying diagnosis or prognosis biomarkers using the logistic regression model is that the number of observations is much smaller than the size of measured biomarkers in most of the genomic studies. Such limitation causes instability in the algorithms used to select gene marker. Regularization methods have been widely used in order to deal with this problem of high dimensionality. For example, Shevade and Keerthi proposed the sparse logistic regression based on the Lasso regularization [1, 2]. Meier et al. investigated logistic regression with group Lasso [3]. Usually, the Lasso type procedures are often called [figure omitted; refer to PDF] -norm type regularization methods. However, [figure omitted; refer to PDF] regularization may yield inconsistent selections when applied to variable selection in some situations [4] and often introduces the extra bias in the estimation [5]. In many genomic studies, we need a sparser solution for interpretation and accurate outcomes, but [figure omitted; refer to PDF] regularization has a gap to meet these requirements. Thus, a further improvement of regularization is urgently required. [figure omitted; refer to PDF] ( [figure omitted; refer to PDF] ) regularization can assuredly generate more sparse and precise solutions than [figure omitted; refer to PDF] regularization. Moreover, [figure omitted; refer to PDF] penalty can be taken as a representative of [figure omitted; refer to PDF] ( [figure omitted; refer to PDF] ) penalty and has demonstrated many attractive properties which do not appear in some [figure omitted; refer to PDF] regularization approaches, such as unbiasedness, sparsity, and oracle properties [6-8].
So far, we observed dense molecular interaction information about the disease-related biological processes and gathered it through databases focused on many aspects of biological systems. For example, BioGRID records collected various biological interactions from more than 43,468 publications [9]. These regulatory relationships are usually represented by a network. Combining these pieces of graphic information extracted from the biological process with an analysis of the gene-expression data had provided useful prior information to detective noise and removes confounding factors from biological data for several classification and regression models [10-14].
Inspired by the aforementioned methods and ideas, here, we define a network-constrained logistic regression model with [figure omitted; refer to PDF] penalty following the framework established by [11], where the predictors are based on the gene-expression data with biologic network knowledge. The proposed model is aimed at identifying some biomarkers and subnetworks regarding diseases. In order to achieve a better prediction, we use an enhanced half thresholding algorithm for [figure omitted; refer to PDF] regularization, which is more efficient than the old half thresholding approach in the literature [6, 15, 16].
The rest of the paper is organized as follows. In Section 2, we proposed a new version of the network-constrained logistic regression model with [figure omitted; refer to PDF] regularization. In Section 3, we presented an enhanced half thresholding method for [figure omitted; refer to PDF] regularization and the corresponding coordinate descent algorithm. In Section 4, we evaluated the performance of our proposed approach on the simulated data and presented the applications of the proposed methods to an analysis of lung cancer data. We concluded the paper with Section 5.
2. [figure omitted; refer to PDF] Penalized Network-Constrained Logistic Regression Model
Generally, assuming that dataset [figure omitted; refer to PDF] has [figure omitted; refer to PDF] samples, [figure omitted; refer to PDF] , where [figure omitted; refer to PDF] is [figure omitted; refer to PDF] th sample with [figure omitted; refer to PDF] genes and [figure omitted; refer to PDF] is the corresponding variable that takes a value of 0 or 1. Define a classifier [figure omitted; refer to PDF] and the logistic regression is defined as [figure omitted; refer to PDF] where [figure omitted; refer to PDF] are the coefficients to be estimated. We can obtain [figure omitted; refer to PDF] by minimizing the log-likelihood function of the logistic regression. Following [11], to combine biological network with an analysis of the gene microarray data, we used a Laplacian constraint approach here. Consider a graph [figure omitted; refer to PDF] , where [figure omitted; refer to PDF] is the set of genes that meet [figure omitted; refer to PDF] explanatory variables and [figure omitted; refer to PDF] is the set of edges. If gene [figure omitted; refer to PDF] and gene [figure omitted; refer to PDF] are connected, then there is an edge between gene [figure omitted; refer to PDF] and gene [figure omitted; refer to PDF] , which is denoted by [figure omitted; refer to PDF] ; else [figure omitted; refer to PDF] . [figure omitted; refer to PDF] denotes the weight of edge [figure omitted; refer to PDF] . The normalized Laplacian matrix [figure omitted; refer to PDF] for [figure omitted; refer to PDF] is defined by [figure omitted; refer to PDF] where [figure omitted; refer to PDF] and [figure omitted; refer to PDF] are the degrees of genes [figure omitted; refer to PDF] and [figure omitted; refer to PDF] , respectively. The degrees of gene [figure omitted; refer to PDF] (or [figure omitted; refer to PDF] ) describe the number of the edges that connected with [figure omitted; refer to PDF] (or [figure omitted; refer to PDF] ). For [figure omitted; refer to PDF] , the network-constrained logistic regression model is presented as [figure omitted; refer to PDF] where the first term in (3) is the log-likelihood function of the logistic model and the second term is a network constraint based on the Laplacian matrix, which induces a smooth solution of [figure omitted; refer to PDF] on the graph.
Directly computing (3) performs poorly for both prediction and biomarker selection purposes when the gene number [figure omitted; refer to PDF] >> the sample size [figure omitted; refer to PDF] . Therefore, the regularization approach is vitally needed. When adding a regularization term to (3), the sparse network-constrained logistic regression can be written as [figure omitted; refer to PDF] where [figure omitted; refer to PDF] is a regularization parameter. In Zhang et al. [13], the authors used Lasso ( [figure omitted; refer to PDF] ) which has the regularization term [figure omitted; refer to PDF] to penalize (4). However, the result of the Lasso type ( [figure omitted; refer to PDF] ) regularization is not good enough for interpretation, especially in genomic research. Besides this, [figure omitted; refer to PDF] regularization is asymptotically biased [17, 18]. To improve the solution's sparsity and its predictive accuracy, we need to think beyond [figure omitted; refer to PDF] regularization to [figure omitted; refer to PDF] penalties. In mathematics, [figure omitted; refer to PDF] type regularization [figure omitted; refer to PDF] with the lower value of [figure omitted; refer to PDF] would lead to better solutions with more sparsity and gives asymptotically unbiased estimates [17]. Moreover, [figure omitted; refer to PDF] penalty can be taken as a representative of [figure omitted; refer to PDF] penalty and has permitted an analytically expressive thresholding representation [6, 7]. Therefore, we proposed a novel [figure omitted; refer to PDF] net approach based on [figure omitted; refer to PDF] regularization to penalize the network-constrained logistic regression model, as shown in [figure omitted; refer to PDF] where [figure omitted; refer to PDF] .
3. A Coordinate Descent Algorithm for the Network-Constrained Logistic Model with the Enhanced [figure omitted; refer to PDF] Thresholding Operator
[figure omitted; refer to PDF] penalty function is nonconvex, which raises numerical challenges in fitting the models. Recently, the coordinate descent algorithms [19] for solving nonconvex regularization models (SCAD [20], MCP [21]) have shown significant efficiency and convergence [22]. Since the computational burden increases only linearly with the feature number [figure omitted; refer to PDF] , the coordinate descent algorithm can be a powerful tool for solving high-dimensional problems. Its standard procedure can be demonstrated as follows: for every coefficient [figure omitted; refer to PDF] , to partially optimize the target function with respect to [figure omitted; refer to PDF] , and fix the remaining elements [figure omitted; refer to PDF] at their most recently updated values. The specific form of updating [figure omitted; refer to PDF] depends on the thresholding operator of the penalty.
In this paper, we present an enhanced [figure omitted; refer to PDF] thresholding operator for the coordinate descent algorithm: [figure omitted; refer to PDF] where [figure omitted; refer to PDF] , [figure omitted; refer to PDF] , [figure omitted; refer to PDF] , and [figure omitted; refer to PDF] as the partial residual for fitting [figure omitted; refer to PDF] .
Remark . This enhanced [figure omitted; refer to PDF] thresholding operator [figure omitted; refer to PDF] outperforms the old [figure omitted; refer to PDF] thresholding [figure omitted; refer to PDF] introduced in [6, 15, 16]. We know that the quantity of the regularization solutions depends seriously on the value of the regularization parameter [figure omitted; refer to PDF] . Based on this enhanced [figure omitted; refer to PDF] thresholding operator, when [figure omitted; refer to PDF] is chosen by some efficient strategies for the parameter tuning, such as cross validation, the convergence of algorithm (6) is proved [7].
The Laplacian matrix [figure omitted; refer to PDF] is nonnegative definite; thus, it can be written as [figure omitted; refer to PDF] by Cholesky decomposition. Following C. Li and H. Li [11] approach, (4) can be expressed as [figure omitted; refer to PDF] where [figure omitted; refer to PDF] , [figure omitted; refer to PDF] , [figure omitted; refer to PDF] , and [figure omitted; refer to PDF] is the regularization parameter and can be expressed as [figure omitted; refer to PDF] .
One-term Taylor series expansion for (7) can be written as [figure omitted; refer to PDF] where [figure omitted; refer to PDF] is the estimated response and [figure omitted; refer to PDF] is the weight for the estimated response. [figure omitted; refer to PDF] is the evaluated value under the current parameters. Thus, we can redefine the partial residual for fitting current [figure omitted; refer to PDF] as [figure omitted; refer to PDF] and [figure omitted; refer to PDF] . The procedure of the coordinate descent algorithm for [figure omitted; refer to PDF] penalized network-constrained logistic model is described as follows.
Algorithm 1 (the coordinate descent algorithm for [figure omitted; refer to PDF] penalized network-constrained logistic model).
We consider the following.
Step 1 . Initialize all [figure omitted; refer to PDF] ( [figure omitted; refer to PDF] ) and [figure omitted; refer to PDF] , [figure omitted; refer to PDF] , and set [figure omitted; refer to PDF] , [figure omitted; refer to PDF] chosen by cross validation.
Step 2 . Calculate [figure omitted; refer to PDF] and [figure omitted; refer to PDF] and approximate the loss function (8) based on the current [figure omitted; refer to PDF] .
Step 3 . Update each [figure omitted; refer to PDF] and cycle over [figure omitted; refer to PDF] , until [figure omitted; refer to PDF] does not change.
Step 3.1 . Compute [figure omitted; refer to PDF] and [figure omitted; refer to PDF] .
Step 3.2 . Update [figure omitted; refer to PDF] .
Step 4 . Let [figure omitted; refer to PDF] , [figure omitted; refer to PDF] .
If [figure omitted; refer to PDF] dose not converge, then repeat Steps 2 and 3.
4. Simulation and Application
4.1. Analyses of Simulated Data
We evaluate the performance of four methods: the network-constrained logistic regression models with [figure omitted; refer to PDF] regularization ( [figure omitted; refer to PDF] net), [figure omitted; refer to PDF] regularization with old thresholding value [figure omitted; refer to PDF] ( [figure omitted; refer to PDF] net) and with the enhanced thresholding value [figure omitted; refer to PDF] (enhanced [figure omitted; refer to PDF] net), and the Elastic net regularization approach (Elastic net). We first simulated the graph structure to mimic gene regulatory network: assuming that the graph consists of 200 independent transcription factors (TFs) and each TF regulates 10 unlike genes, so there are a total of 2200 variables, [figure omitted; refer to PDF] , [figure omitted; refer to PDF] . The training and the independent test data sets include the sample sizes of 100, respectively. Each TF [figure omitted; refer to PDF] and its regulated genes [figure omitted; refer to PDF] were generated by the normal distribution [figure omitted; refer to PDF] . We set the correlation rate between [figure omitted; refer to PDF] and its regulated gene [figure omitted; refer to PDF] as 0.75, [figure omitted; refer to PDF] . The binary responder [figure omitted; refer to PDF] ( [figure omitted; refer to PDF] ), which is associated with the matrix [figure omitted; refer to PDF] of TFs and their regulated genes, is calculated based on the following formula and rule: [figure omitted; refer to PDF] where [figure omitted; refer to PDF] , [figure omitted; refer to PDF] for Model 1, and [figure omitted; refer to PDF] .
Model 2 was defined similar to Model 1, except that we considered the case when the TF can have positive and negative effects on its regulated genes at the same time: [figure omitted; refer to PDF]
In these two models, the 10-fold cross validation approach was conducted on the training datasets to tune the regularization parameters of the enhanced [figure omitted; refer to PDF] net, [figure omitted; refer to PDF] net, and [figure omitted; refer to PDF] net. Both penalized parameters for [figure omitted; refer to PDF] and ridge regularization in the Elastic net were tuned by the 10-fold cross validation on the two-dimensional parameter surfaces. We repeated the simulations over 100 times and then computed the misclassification error, the sensitivity, and the specificity averagely for each net model on the test datasets.
Table 1 summarizes the simulation results from each regularization net model. In general, our proposed enhanced [figure omitted; refer to PDF] net model achieved the smallest misclassification errors in Models 1 (9.22%) and 2 (10.76%) compared with the other regularization methods including the old [figure omitted; refer to PDF] thresholding method (9.85% for Model 1 and 10.83% for Model 2), [figure omitted; refer to PDF] net (11.81% for Model 1 and 13.21% for Model 2), and the Elastic net (13.12% for Model 1 and 14.14% for Model 2). Meanwhile, the enhanced [figure omitted; refer to PDF] net resulted in the highest sensitivity in Model 1 (98.5%) compared with the other methods. Moreover, the enhanced [figure omitted; refer to PDF] net obtained the best specificity in Model 2 (98.7%) amongst the other approaches. To sum up, the enhanced [figure omitted; refer to PDF] net outperforms the other three algorithms in terms of prediction accuracy, sensitivity, and specificity.
Table 1: Simulation results of the enhanced [figure omitted; refer to PDF] net, [figure omitted; refer to PDF] net, [figure omitted; refer to PDF] net, and Elastic net, respectively.
Model | Misclassification errors (%) | Sensitivity (%) | Specificity (%) | |||||||||
Eh_ [figure omitted; refer to PDF] | [figure omitted; refer to PDF] | [figure omitted; refer to PDF] | Elastic | Eh_ [figure omitted; refer to PDF] | [figure omitted; refer to PDF] | [figure omitted; refer to PDF] | Elastic | Eh_ [figure omitted; refer to PDF] | [figure omitted; refer to PDF] | [figure omitted; refer to PDF] | Elastic | |
1 | 9.22 | 9.85 | 11.81 | 13.12 | 0.985 | 0.971 | 0.968 | 0.873 | 0.969 | 0.970 | 0.962 | 0.981 |
(0.36) | (0.31) | (0.41) | (0.12) | (0.00) | (0.00) | (0.02) | (0.00) | (0.00) | (0.01) | (0.01) | (0.00) | |
| ||||||||||||
2 | 10.76 | 10.83 | 13.21 | 14.14 | 0.939 | 0.939 | 0.943 | 0.835 | 0.987 | 0.981 | 0.987 | 0.980 |
(0.33) | (0.36) | (0.24) | (0.23) | (0.00) | (0.00) | (0.01) | (0.00) | (0.02) | (0.01) | (0.01) | (0.00) |
Simulation results (averaged over 100 runs) for comparison of misclassification errors, sensitivity, and specificity used the enhanced [figure omitted; refer to PDF] net, [figure omitted; refer to PDF] net, [figure omitted; refer to PDF] net, and the Elastic net, respectively. The standard errors are given in parentheses.
4.2. Analysis of Lung Cancer
In this section, we merged the protein-protein interaction (PPI) network (see http://thebiogrid.org/) with a lung cancer (LC) gene-expression dataset [23] to demonstrate the performance of our proposed enhanced [figure omitted; refer to PDF] net method. The gene-expression dataset contains the expression profiles of 22284 genes for 107 patients, in which 58 had lung cancer. To test the generalization ability of the proposed method, we divided the dataset into the training set (sample size [figure omitted; refer to PDF] ; 38 LC, 32 non-LC) which covered 2/3 samples of the dataset and the test set (sample size [figure omitted; refer to PDF] ; 20 LC, 17 non-LC) which covered the other 1/3 specimens of the dataset. The 10-fold cross validation approach was conducted on the training dataset to tune the regularization parameters. By combining the gene-expression data with the PPI network, the final PPI network includes 8619 genes and 28293 edges.
Figures 1-4 display the solution paths of the four regularization net methods for the LC dataset in one sample run. Here, [figure omitted; refer to PDF] -axis displays the values of the running lambda (the running lambda of [figure omitted; refer to PDF] penalty in the Elastic net approach), and [figure omitted; refer to PDF] -axis at the top (degrees of freedom) means the number of nonzero coefficients of beta. [figure omitted; refer to PDF] -axis is the values of the coefficients beta which measure the gene importance. The predictive model builds from the training set and then tests its predictive performance on the test set. The detailed results were represented in Table 2.
Table 2: The results of the enhanced [figure omitted; refer to PDF] net, [figure omitted; refer to PDF] net, [figure omitted; refer to PDF] net, and Elastic net on LC dataset, respectively.
| Selected genes | Connected genes | Connected edges | Cross validation error | Test error |
Eh_ [figure omitted; refer to PDF] net | 171 | 54 | 41 | 6/70 | 5/37 |
[figure omitted; refer to PDF] net | 193 | 61 | 47 | 6/70 | 6/37 |
[figure omitted; refer to PDF] net | 500 | 150 | 121 | 7/70 | 6/37 |
[figure omitted; refer to PDF] | 636 | 337 | 510 | 6/70 | 6/37 |
Results of analysis of LC gene expression dataset by four procedures, including the number of genes selected, the number of linked PPI network genes, the number of linked PPI network edges, the CV error, and test errors.
Figure 1: The solution paths of the enhanced [figure omitted; refer to PDF] net for the lung cancer dataset in one sample run.
[figure omitted; refer to PDF]
Figure 2: The solution paths of [figure omitted; refer to PDF] net for the lung cancer dataset in one sample run.
[figure omitted; refer to PDF]
Figure 3: The solution paths of [figure omitted; refer to PDF] net for the lung cancer dataset in one sample run.
[figure omitted; refer to PDF]
Figure 4: The solution paths of the Elastic net for the lung cancer dataset in one sample run.
[figure omitted; refer to PDF]
As shown in Table 2, the enhanced [figure omitted; refer to PDF] net selected the fewest number of genes and edges compared to [figure omitted; refer to PDF] net, [figure omitted; refer to PDF] net, and the Elastic net. Meanwhile, the predictive performance of the enhanced [figure omitted; refer to PDF] net outperforms the other three regularization net algorithms.
To further evaluate the performance of the enhanced [figure omitted; refer to PDF] net procedure, we report its capacity of identifying the biomarkers related to lung cancer. NK2 homeobox 1 (Nkx2-1) protein regulates transcription of genes specific for lung. It is used as a biomarker to determine lung cancer in anatomic pathology. It also has a critical role in maintaining lung tumor cells [24, 25]. Epidermal growth factor receptor (EGFR) is known to play a key role in cell proliferation and apoptosis. EGFR overexpression and activity could result in tumor growth and progression [26] and somatic mutations within the tyrosine kinase domain of EGFR, which have been identified in a subset of lung adenocarcinoma [27, 28]. The enhanced [figure omitted; refer to PDF] net (Figure 5) and [figure omitted; refer to PDF] net successfully identified these two important biomarkers for LC. However, neither [figure omitted; refer to PDF] net nor the Elastic net selected them both.
Figure 5: Subnetworks identified by the enhanced [figure omitted; refer to PDF] net for lung cancer datasets (only those genes that are linked on the PPI network are plotted).
[figure omitted; refer to PDF]
Except to identify these two significant biomarkers (EGFR and Nkx2-1), the enhanced [figure omitted; refer to PDF] net also selected several pathways that were associated with lung cancer. For example, one of the subnetworks includes genes involving molecular proliferation (e.g., genes ARF4, EGFR, DCN, BRCA1, and ITIH5). As these gene express significantly and continuously, it promotes lung cancer progression. On the other hand, this group is linked to ENO1. We are unable to get a clear testimony to sustain this relationship by looking at PPI database. However, a recent report [29] has demonstrated that ENO1 is the promising biomarker that may provide more diagnostic efficacy for lung cancer. This link implies a functional relationship and suggests the important role of ENO1 in lung cancer.
All these results reveal that the enhanced [figure omitted; refer to PDF] net is more reliable than [figure omitted; refer to PDF] net, [figure omitted; refer to PDF] net, and the Elastic net approaches for selecting key markers from high-dimensional genomic data. Another advantage of our proposed method is that it has the ability to recognize novel and potential relationships with biologic significance. It is mentionable that our proposed method is inclined to identify fewer but more informative genes (or edges) than [figure omitted; refer to PDF] net and the Elastic net approaches in genomic data and that means the proposed method has allowed the researcher to more easily concentrate on the key targets for functional studies or downstream applications.
5. Conclusions
In biological molecular research, especially for cancer, the analysis of combining biological pathway information with gene-expression data may play an important role to search for new targets for drug design. In this paper, we use the enhanced [figure omitted; refer to PDF] solver to penalized network-constrained logistic regression model to integrate lung cancer gene-expression with protein-to-protein interaction network. We develop the corresponding coordinate descent algorithm as a novel biomarker selection approach. This algorithm is extremely fast and easy to implement. Both simulation and real genomic data studies showed that the enhanced [figure omitted; refer to PDF] net is a ranking procedure compared with [figure omitted; refer to PDF] net (using the old thresholding operator), [figure omitted; refer to PDF] net, and the Elastic net in the selection of biomarker and subnetwork.
We successfully identified several important clinical biomarkers and subnetwork that are driving lung cancer. The proposed method has provided new information to investigators in biological studies and can be the efficient tool for identifying cancer related biomarker and subnetwork.
Acknowledgment
This work was supported by the Macau Science and Technology Development Funds (Grant no. 099/2013/A3) of Macau SAR of China.
Conflict of Interests
The authors declare no conflict of interests.
[1] S. K. Shevade, S. S. Keerthi, "A simple and efficient algorithm for gene selection using sparse logistic regression," Bioinformatics , vol. 19, no. 17, pp. 2246-2253, 2003.
[2] R. Tibshirani, "Regression shrinkage and selection via the lasso," Journal of the Royal Statistical Society Series B: Methodological , vol. 58, no. 1, pp. 267-288, 1996.
[3] L. Meier, S. van de Geer, P. Bühlmann, "The group Lasso for logistic regression," Journal of the Royal Statistical Society. Series B. Statistical Methodology , vol. 70, no. 1, pp. 53-71, 2008.
[4] H. Zou, "The adaptive lasso and its oracle properties," Journal of the American Statistical Association , vol. 101, no. 476, pp. 1418-1429, 2006.
[5] N. Meinshausen, B. Yu, "Lasso-type recovery of sparse representations for high-dimensional data," The Annals of Statistics , vol. 37, no. 1, pp. 246-270, 2009.
[6] Z. B. Xu, H. Zhang, Y. Wang, X. Y. Chang, Y. Liang, "L1/2 regularization," Science in China Series F , vol. 40, no. 3, pp. 1-11, 2010.
[7] Z. Xu, X. Chang, F. Xu, H. Zhang, " L 1/2 regularization: a thresholding representation theory and a fast solver," IEEE Transactions on Neural Networks and Learning Systems , vol. 23, no. 7, pp. 1013-1027, 2012.
[8] J. Zeng, S. Lin, Y. Wang, " L 1 / 2 regularization: convergence of iterative half thresholding algorithm," IEEE Transactions on Signal Processing , vol. 62, no. 9, pp. 2317-2329, 2014.
[9] C. Stark, B.-J. Breitkreutz, T. Reguly, L. Boucher, A. Breitkreutz, M. Tyers, "BioGRID: a general repository for interaction datasets,", supplement 1 Nucleic acids research. , vol. 34, pp. D535-D539, 2006.
[10] H.-Y. Chuang, E. Lee, Y.-T. Liu, D. Lee, T. Ideker, "Network-based classification of breast cancer metastasis," Molecular Systems Biology , vol. 3, article 140, 2007.
[11] C. Li, H. Li, "Network-constrained regularization and variable selection for analysis of genomic data," Bioinformatics , vol. 24, no. 9, pp. 1175-1182, 2008.
[12] Z. Tian, T. Hwang, R. Kuang, "A hypergraph-based learning algorithm for classifying gene expression and arrayCGH data with prior knowledge," Bioinformatics , vol. 25, no. 21, pp. 2831-2838, 2009.
[13] W. Zhang, Y.-W. Wan, G. I. Allen, K. Pang, M. L. Anderson, Z. Liu, "Molecular pathway identification using biological network-regularized logistic models," BMC Genomics , vol. 14, no. 8, article S7, 2013.
[14] H. Sun, W. Lin, R. Feng, H. Li, "Network-regularized high dimensional Cox regression for Analysis of Genomic data," Statistica Sinica , vol. 24, pp. 1433-1459, 2014.
[15] Y. Liang, C. Liu, X.-Z. Luan, K.-S. Leung, T.-M. Chan, Z.-B. Xu, H. Zhang, "Sparse logistic regression with a L1/2 penalty for gene selection in cancer classification," BMC Bioinformatics , vol. 14, no. 1, article 198, 2013.
[16] C. Liu, Y. Liang, X.-Z. Luan, K.-S. Leung, T.-M. Chan, Z.-B. Xu, H. Zhang, "The L1/2 regularization method for variable selection in the Cox model," Applied Soft Computing Journal , vol. 14, pp. 498-503, 2014.
[17] K. Knight, W. Fu, "Asymptotics for Lasso-type estimators," Annals of Statistics , vol. 28, no. 5, pp. 1356-1378, 2000.
[18] D. Malioutov, M. Çetin, A. S. Willsky, "A sparse signal reconstruction perspective for source localization with sensor arrays," IEEE Transactions on Signal Processing , vol. 53, no. 8, pp. 3010-3022, 2005.
[19] J. Friedman, T. Hastie, R. Tibshirani, "Regularization paths for generalized linear models via coordinate descent," Journal of Statistical Software , vol. 33, no. 1, pp. 1-22, 2010.
[20] J. Fan, R. Li, "Variable selection via nonconcave penalized likelihood and its oracle properties," Journal of the American Statistical Association , vol. 96, no. 456, pp. 1348-1360, 2001.
[21] C.-H. Zhang, "Nearly unbiased variable selection under minimax concave penalty," The Annals of Statistics , vol. 38, no. 2, pp. 894-942, 2010.
[22] P. Breheny, J. Huang, "Coordinate descent algorithms for nonconvex penalized regression, with applications to biological feature selection," Annals of Applied Statistics , vol. 5, no. 1, pp. 232-253, 2011.
[23] M. T. Landi, T. Dracheva, M. Rotunno, J. D. Figueroa, H. Liu, A. Dasgupta, F. E. Mann, J. Fukuoka, M. Hames, A. W. Bergen, S. E. Murphy, P. Yang, A. C. Pesatori, D. Consonni, P. A. Bertazzi, S. Wacholder, J. H. Shih, N. E. Caporaso, J. Jen, "Gene expression signature of cigarette smoking and its role in lung adenocarcinoma development and survival," PLoS ONE , vol. 3, no. 2, 2008.
[24] Y. Li, K. Eggermont, V. Vanslembrouck, C. M. Verfaillie, "NKX2-1 activation by SMAD2 signaling after definitive endoderm differentiation in human embryonic stem cell," Stem Cells and Development , vol. 22, no. 9, pp. 1433-1442, 2013.
[25] B. A. Weir, M. S. Woo, G. Getz, "Characterizing the cancer genome in lung adenocarcinoma," Nature , vol. 450, no. 7171, pp. 893-898, 2007.
[26] H. Zhang, A. Berezov, Q. Wang, G. Zhang, J. Drebin, R. Murali, M. I. Greene, "ErbB receptors: from oncogenes to targeted cancer therapies," The Journal of Clinical Investigation , vol. 117, no. 8, pp. 2051-2058, 2007.
[27] J. G. Paez, P. A. Janne, J. C. Lee, S. Tracy, H. Greulich, S. Gabriel, P. Herman, F. J. Kaye, N. Lindeman, T. J. Boggon, K. Naoki, H. Sasaki, Y. Fujii, M. J. Eck, W. R. Sellers, B. E. Johnson, M. Meyerson, "EGFR mutations in lung cancer: correlation with clinical response to gefitinib therapy," Science , vol. 304, no. 5676, pp. 1497-1500, 2004.
[28] S. R. Chowdhuri, L. Xi, T. H.-T. Pham, J. Hanson, J. Rodriguez-Canales, A. Berman, A. Rajan, G. Giaccone, M. Emmert-Buck, M. Raffeld, A. C. Filie, "EGFR and KRAS mutation analysis in cytologic samples of lung adenocarcinoma enabled by laser capture microdissection," Modern Pathology , vol. 25, no. 4, pp. 548-555, 2012.
[29] L. Yu, J. Shen, K. Mannoor, M. Guarnera, F. Jiang, "Identification of ENO1 as a potential sputum biomarker for early stage lung cancer by shotgun proteomics," Clinical Lung Cancer , vol. 15, no. 5, pp. 372-378, 2014.
You have requested "on-the-fly" machine translation of selected content from our databases. This functionality is provided solely for your convenience and is in no way intended to replace human translation. Show full disclaimer
Neither ProQuest nor its licensors make any representations or warranties with respect to the translations. The translations are automatically generated "AS IS" and "AS AVAILABLE" and are not retained in our systems. PROQUEST AND ITS LICENSORS SPECIFICALLY DISCLAIM ANY AND ALL EXPRESS OR IMPLIED WARRANTIES, INCLUDING WITHOUT LIMITATION, ANY WARRANTIES FOR AVAILABILITY, ACCURACY, TIMELINESS, COMPLETENESS, NON-INFRINGMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Your use of the translations is subject to all use restrictions contained in your Electronic Products License Agreement and by using the translation functionality you agree to forgo any and all claims against ProQuest or its licensors for your use of the translation functionality and any output derived there from. Hide full disclaimer
Copyright © 2015 Hai-Hui Huang et al. Hai-Hui Huang et al. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.
Abstract
Identifying biomarker and signaling pathway is a critical step in genomic studies, in which the regularization method is a widely used feature extraction approach. However, most of the regularizers are based on [subscript]L1[/subscript] -norm and their results are not good enough for sparsity and interpretation and are asymptotically biased, especially in genomic research. Recently, we gained a large amount of molecular interaction information about the disease-related biological processes and gathered them through various databases, which focused on many aspects of biological systems. In this paper, we use an enhanced [subscript]L1/2[/subscript] penalized solver to penalize network-constrained logistic regression model called an enhanced [subscript]L1/2[/subscript] net, where the predictors are based on gene-expression data with biologic network knowledge. Extensive simulation studies showed that our proposed approach outperforms [subscript]L1[/subscript] regularization, the old [subscript]L1/2[/subscript] penalized solver, and the Elastic net approaches in terms of classification accuracy and stability. Furthermore, we applied our method for lung cancer data analysis and found that our method achieves higher predictive accuracy than [subscript]L1[/subscript] regularization, the old [subscript]L1/2[/subscript] penalized solver, and the Elastic net approaches, while fewer but informative biomarkers and pathways are selected.
You have requested "on-the-fly" machine translation of selected content from our databases. This functionality is provided solely for your convenience and is in no way intended to replace human translation. Show full disclaimer
Neither ProQuest nor its licensors make any representations or warranties with respect to the translations. The translations are automatically generated "AS IS" and "AS AVAILABLE" and are not retained in our systems. PROQUEST AND ITS LICENSORS SPECIFICALLY DISCLAIM ANY AND ALL EXPRESS OR IMPLIED WARRANTIES, INCLUDING WITHOUT LIMITATION, ANY WARRANTIES FOR AVAILABILITY, ACCURACY, TIMELINESS, COMPLETENESS, NON-INFRINGMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Your use of the translations is subject to all use restrictions contained in your Electronic Products License Agreement and by using the translation functionality you agree to forgo any and all claims against ProQuest or its licensors for your use of the translation functionality and any output derived there from. Hide full disclaimer