This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.
1. Introduction
Protein post-translational modification (PTM) refers to the chemical interaction occurring prior to protein biosynthesis and after mRNAs are translated into polypeptide chains. PTM has different categories and is very prevalent in the cells. More than 450 categories of PTMs were discovered to date, such as phosphorylation, methylation, and acetylation [1–3]. PTM increases diversity of protein structures and functions, viewed as one of most regulating mechanisms in the cellular process. Lysine succinylation is a type of protein TPMs, in which a succinyl group (-CO-CH2-CH2-CO2H) is attached to lysine residue of proteins [4]. Succinylation is reversible, dynamic, and evolutionarily conserved, widely existing in the prokaryote and the eukaryotes cells [5, 6]. The succinylation of proteins induces shift in the charge and the structural alteration and thus would yield effects on functions of proteins [6]. Growing evidences also showed aberrant succinylations were involved in the pathogenesis of some diseases including cancers [7], metabolism disease [8, 9], and nervous system diseases [10]. Thus, identifying succinylation sites and understanding its mechanism are crucial to develop drugs for related diseases.
Identifying succinylation sites has two main routes: experimental and computational methods. The experimental methods were represented by mass spectrometry, which contributed to the validation of succinylation and collection of first-hand data. On the other hand, the experimental methods are labor-intensive and time-consuming without assist of the computational methods. The computational methods are based on data yielded by the experimental methods and build machine learning-based models to predict new succinylations. Therefore, identifying succinylation is a cyclic iterative process from experiment to computation and again from computation to experiment. We focused on the computational methods to predict succinylation. In the past decades, more than ten computational methods have been developed for identifying succinylation [11–29]. Most of these computational methods extracted features directly from protein sequences, which were subsequently used for training model. For example, Zhao et al. [11] used the auto-correlation functions, the group weight-based encoding, the normalized van der Waals volume, and the position weight amino acid composition. Kao et al. [25] exploited the amino acid composition and informative
2. Data
All the succinylated proteins were downloaded from the PLMD (Protein Lysine Modifications Database) database which is dedicated to specifically collect protein lysine modification [30–32]. The PLMD has evolved to version 3.0, housing 284780 modification events in 53501 proteins for 20 types of lysine modification. We extracted 6377 proteins containing 18593 succinylation sites. To remove dependency of the proposed method on the homology, we used the software CD-Hit [33, 34] to cluster 6377 protein sequences. The sequence identify cut-off was set to 0.4, and we obtained 3560 protein sequences, of which any two kept sequence similarity less than 0.4. We randomly divided these 3560 proteins into the training and the testing samples at the ratio of training to testing 4 : 1, resulting in 712 testing and 2848 training sequences. For each protein sequence, we extracted all the peptides which centered the lysine residue with 15 amino acid residues in the downstream/upstream of it. For peptides less than 15 amino acid residues, we prefixed or suffixed “
3. Method
As shown in Figure 1, the proposed deep learning network consisted mainly of embedding, 1D convolution, pooling, bidirectional LTSM, dropout, flatten, and fully connected layers. Peptides with 31 amino acid residues were entered to the embedding layer and were translated into vectors with shape of (31, 64). Then, two different network structures, respectively, took the embedding as input, and their outputs were concatenated as input to the fully connected layer. One structure was the convolution neural network, and another was the bidirectional LSTM neural network. The final output was a neuron representing probability of belonging to the positive sample. The parameters and the shape of output of each layers in the deep neural network are listed in Table 1. The total number of trainable parameters is 336,897.
[figure omitted; refer to PDF]
Table 1
Number of parameters and shape of output in the LSTMCNNsucc.
| Layers | Parameters | Output |
| Embedding | 1472 | (None, 31, 64) |
| Bidirectional LSTM | 197632 | (None, 31, 256) |
| Dropout | 0 | (None, 31, 256) |
| Flatten | 0 | (None, 7936) |
| 1D convolution | 10272 | (None, 27, 32) |
| Pooling | 0 | (None, 32) |
| Dense (16) | 127504 | (None, 16) |
| Dense (1) | 17 | (None, 1) |
3.1. Embedding Layer
Most machine learning-based methods for predicting protein post-translational modification generally required an encoding step which translated sequences into vector representation. For example, the frequently used encoding schemes included position specific scoring matrix [35], amino acid composition, composition of
3.2. 1D Convolution Layer
The convolution neural network (CNN) proposed by LeCun et al. [39, 40] is a feed forward network. Compared with the conventional neural network, the CNN has two notable properties: local connectivity and parameter sharing. The local connectivity lies that two neighboring layers are not fully connected but locally connected. That is to say, the neuron in a layer is not connected to all neurons in the neighboring layers. The CNN implemented the parameter sharing via the filter (also called convolution kernel). The filter slides on the image and convoluted with all sections in image. The filter is shared by the image. In the last ten years, many deep convolution neural networks such as AlexNet [41], VGG [42], GoogleNet [43], and ResNet [44] have been proposed and applied to computer vision. The CNN achieved significant advance in terms of classification error in comparison with the previous deep neural network. The convolution is of 1-dimension, 2-dimension, or more than 2 dimensions. Here, we used 1D convolution. Suppose a discrete sequence was
where
3.3. Pooling Layer
The pooling operation firstly appeared in the AlexNet [41] and is increasingly becoming one of components of the deep CNN architecture. The pooling operation has such categories as max pooling, min pooling, and mean pooling. The role of pooling operation included removal of redundancy information and reduction of overfitting. Here, we used the max pooling operation. Given an
3.4. Bidirectional LSTM Layer
Recurrent neural network (RNN) [45, 46] is a different framework of neural network from multiple layer perception. The RNN shares weights and is especially suitable to the field of sequence analysis such as language translation and semantic understanding. An unfolded RNN model was shown in Figure 2(a). The hidden state
[figures omitted; refer to PDF]
The previous RNN was forward. The output at the time step
The output at the time step
3.5. Dropout Layer
The deep neural network is prone to lead to overfitting when the number of training samples was too less. To deal with this issue, Hinton et al. [49] proposed the dropout concept. Due to its effect and efficiency, the dropout is increasingly becoming the frequently used trick in the deep learning area [41, 50–53]. The neurons were dropped out at a certain rate of dropout, and parameters of only preserved neurons were updated in the training stage, while all the neurons were used in the predicting stage.
3.6. Flatten Layer and Fully Connected Layer
The role of flatten layer was only to convert the data into one-dimension and then facilitated connection of the fully connected layer. No parameters were trainable in the flatten layer. The fully connected layer was similar to hidden layer in the MLP, each neuron connected to the neurons in the preceding layer.
4. Metrics
We adopted to evaluate the predicted result these frequently used metrics in the binary classification questions such as sensitivity (SN), specificity (SP), accuracy (ACC), and Matthews correlation coefficient (MCC), which were defined by
5. Results
Table 2 showed the predicting performance of the trained model on the 712 testing sequences. Although more than ten approaches or tools for predicting succinylation have been proposed in the past ten years, either they did not provide online predicting server or the web server could not work. We compared the proposed method to three methods whose web predicting server still can work [28]: SuccinSite [15], iSuc-PseAAC [12], and DeepSuccinylSite [29]. 712 testing sequences were used to examine three approaches. Among 712 testing sequences, at least 225 sequences repeated in the training set of the SuccinSite, and at least 223 repeated in the training set of DeepSuccinylSite. These minus 225 sequences were used to examine the SuccinSite and these minus 223 sequences to test the DeepSuccinylSite. iSuc-PseAAC [12] obtained best SP and best ACC but worst SN and worst MCC. The SuccinSite [15] reached better SP and better ACC but worse MCC and worse SN. The iSuc-PseAAC [12] and the SuccinSite [15] were in favor of predicting the negative samples. The DeepSuccinylSite [29] was better than the LSTMCNNsucc in terms of SN, worse than the LSTMCNNsucc in terms of sp. The overall performance of the LSTMCNNsucc was slightly better than that of the DeepSuccinylSite.
Table 2
Comparison with state of the art methods.
| Method | SN | SP | ACC | MCC |
| LSTMCNNsucc | 0.5916 | 0.7957 | 0.7789 | 0.2508 |
| SuccinSite [15] | 0.3977 | 0.8635 | 0.8272 | 0.1925 |
| iSuc-PseAAC [12] | 0.1258 | 0.8929 | 0.8296 | 0.0165 |
| DeepSuccinylSite [29] | 0.7438 | 0.6879 | 0.6923 | 0.2438 |
5.1. Functional Analysis
We used the statistical over-representation test of gene list analysis in the PANTHER classification system [54, 55] to perform function enrichment analysis of the succinylated proteins. The significant biological process, the molecular function, and the cellular component terms (
[figures omitted; refer to PDF]
We also performed enrichment analysis of Kyoto Encyclopedia of Genes and Genomes (KEGG) pathway by functional annotation in the DAVID tool [56, 57] to investigate in which pathway the succinylated proteins were involved. The statistically significant KEGG terms (
Table 3
Significant KEGG pathway terms.
| Species | KEGG terms | Benjamini |
| E. coli | Metabolic pathways | 3.30 |
| Biosynthesis of amino acids | 1.00 | |
| Biosynthesis of secondary metabolites | 2.40 | |
| Biosynthesis of antibiotics | 7.40 | |
| Lysine biosynthesis | 3.30 | |
| H. sapiens | Biosynthesis of antibiotics | 3.70 |
| Metabolic pathways | 2.80 | |
| Ribosome | 3.40 | |
| Valine, leucine, and isoleucine degradation | 1.30 | |
| Carbon metabolism | 6.20 | |
| Oxidative phosphorylation | 1.10 | |
| Parkinson’s disease | 2.60 | |
| Citrate cycle (TCA cycle) | 1.00 | |
| Huntington’s disease | 4.10 | |
| Alzheimer’s disease | 7.80 | |
| Aminoacyl-tRNA biosynthesis | 1.00 | |
| Butanoate metabolism | 3.40 | |
| Proteasome | 8.20 | |
| M. musculus | Metabolic pathways | 6.20 |
| Parkinson’s disease | 8.50 | |
| Oxidative phosphorylation | 3.40 | |
| Nonalcoholic fatty liver disease (NAFLD) | 1.00 | |
| Huntington’s disease | 2.80 | |
| Alzheimer’s disease | 1.40 | |
| Ribosome | 3.30 | |
| Peroxisome | 1.80 | |
| Glycine, serine, and threonine metabolism | 1.50 | |
| Pyruvate metabolism | 9.00 | |
| Propanoate metabolism | 2.40 | |
| Valine, leucine, and isoleucine degradation | 1.90 | |
| Glyoxylate and dicarboxylate metabolism | 3.10 | |
| Biosynthesis of antibiotics | 5.60 | |
| M. tuberculosis | Metabolic pathways | 1.00 |
| Microbial metabolism in diverse environments | 2.50 | |
| Biosynthesis of antibiotics | 4.40 | |
| Biosynthesis of secondary metabolites | 1.00 | |
| Propanoate metabolism | 1.00 | |
| S. cerevisiae | Metabolic pathways | 5.20 |
| Biosynthesis of amino acids | 3.30 | |
| 2-Oxocarboxylic acid metabolism | 7.90 | |
| Biosynthesis of antibiotics | 3.50 | |
| Oxidative phosphorylation | 3.50 | |
5.2. LSTMCNNsucc Web Server
We built a web server of the proposed LSTMCNNsucc at http://8.129.111.5/. Users either directly input protein sequences in a fasta format or upload a file of fasta format to perform prediction. When both protein sequences and files were submitted, the file was given to priority of prediction.
6. Conclusion
We presented a bidirectional LSTM and CNN-based deep learning method for predicting succinylation sites. The method absorbed semantic relationship hidden in the succinylation sequences, outperforming state-of-the-art method. The functions of succinylation proteins were conserved to a certain extent across species but also were species-specific. We also implemented the proposed method into a user-friendly web server which is available at http://8.129.111.5/.
Acknowledgments
This work is supported by the National Natural Science Foundation of China (11871061, 61672356), by the Scientific Research Fund of Hunan Provincial Education Department (18A253), by the open project of Hunan Key Laboratory for Computation and Simulation in Science and Engineering (2019LCESE03), and by Shaoyang University Graduate Research Innovation Project (CX2020SY060).
[1] E. S. Witze, W. M. Old, K. A. Resing, N. G. Ahn, "Mapping protein post-translational modifications with mass spectrometry," Nature Methods, vol. 4 no. 10, pp. 798-806, DOI: 10.1038/nmeth1100, 2007.
[2] B. S. Sharma, "Post-translational modifications (PTMs), from a cancer perspective: an overview," Oncogen Journal, vol. 2 no. 3, 2019.
[3] G. Huang, X. Li, "A review of computational identification of protein post-translational modifications," Mini-Reviews in Organic Chemistry, vol. 12 no. 6, pp. 468-480, DOI: 10.2174/1570193X13666151218191044, 2015.
[4] Z. Zhang, M. Tan, Z. Xie, L. Dai, Y. Chen, Y. Zhao, "Identification of lysine succinylation as a new post-translational modification," Nature Chemical Biology, vol. 7 no. 1, pp. 58-63, DOI: 10.1038/nchembio.495, 2011.
[5] J. Du, Y. Zhou, X. Su, J. J. Yu, S. Khan, H. Jiang, J. Kim, J. Woo, J. H. Kim, B. H. Choi, "Sirt5 is a NAD-dependent protein lysine demalonylase and desuccinylase," Science, vol. 334 no. 6057, pp. 806-809, DOI: 10.1126/science.1207861, 2011.
[6] A. Sreedhar, E. K. Wiese, T. Hitosugi, "Enzymatic and metabolic regulation of lysine succinylation," Genes & Diseases, vol. 7 no. 2, pp. 166-171, DOI: 10.1016/j.gendis.2019.09.011, 2020.
[7] Y. Xiangyun, N. Xiaomin, "Desuccinylation of pyruvate kinase M2 by SIRT5 contributes to antioxidant response and tumor growth," Oncotarget, vol. 8 no. 4, pp. 6984-6993, DOI: 10.18632/oncotarget.14346, 2017.
[8] M. Yang, Y. Wang, Y. Chen, Z. Cheng, J. Gu, J. Deng, L. Bi, C. Chen, R. Mo, X. Wang, "Succinylome analysis reveals the involvement of lysine succinylation in metabolism in pathogenic Mycobacterium tuberculosis," Molecular & Cellular Proteomics, vol. 14 no. 4, pp. 796-811, DOI: 10.1074/mcp.M114.045922, 2015.
[9] Y. Yang, G. E. Gibson, "Succinylation links metabolism to protein functions," Neurochemical Research, vol. 44 no. 10, pp. 2346-2359, DOI: 10.1007/s11064-019-02780-x, 2019.
[10] G. E. Gibson, H. Xu, H. L. Chen, W. Chen, T. T. Denton, S. Zhang, "Alpha-ketoglutarate dehydrogenase complex-dependent succinylation of proteins in neurons and neuronal cell lines," Journal of Neurochemistry, vol. 134 no. 1, pp. 86-96, DOI: 10.1111/jnc.13096, 2015.
[11] X. Zhao, Q. Ning, H. Chai, Z. Ma, "Accurate in silico identification of protein succinylation sites using an iterative semi-supervised learning technique," Journal of Theoretical Biology, vol. 374, pp. 60-65, DOI: 10.1016/j.jtbi.2015.03.029, 2015.
[12] Y. Xu, Y. X. Ding, J. Ding, Y. H. Lei, L. Y. Wu, N. Y. Deng, "iSuc-PseAAC: predicting lysine succinylation in proteins by incorporating peptide position-specific propensity," Scientific Reports, vol. 5 no. 1, article 10184,DOI: 10.1038/srep10184, 2015.
[13] J. Jia, Z. Liu, X. Xiao, B. Liu, K. C. Chou, "iSuc-PseOpt: identifying lysine succinylation sites in proteins by incorporating sequence-coupling effects into pseudo components and optimizing imbalanced training dataset," Analytical Biochemistry, vol. 497, pp. 48-56, DOI: 10.1016/j.ab.2015.12.009, 2016.
[14] H. D. Xu, S. P. Shi, P. P. Wen, J. D. Qiu, "SuccFind: a novel succinylation sites online prediction tool via enhanced characteristic strategy," Bioinformatics, vol. 31 no. 23, pp. 3748-3750, DOI: 10.1093/bioinformatics/btv439, 2015.
[15] M. M. Hasan, S. Yang, Y. Zhou, M. N. Mollah, "SuccinSite: a computational tool for the prediction of protein succinylation sites by exploiting the amino acid patterns and properties," Molecular BioSystems, vol. 12 no. 3, pp. 786-795, DOI: 10.1039/C5MB00853K, 2016.
[16] H. Ai, R. Wu, L. Zhang, X. Wu, J. Ma, H. Hu, L. Huang, W. Chen, J. Zhao, H. Liu, "pSuc-PseRat: predicting lysine succinylation in proteins by exploiting the ratios of sequence coupling and properties," Journal of Computational Biology, vol. 24 no. 10, pp. 1050-1059, DOI: 10.1089/cmb.2016.0206, 2017.
[17] A. Dehzangi, Y. López, S. P. Lal, G. Taherzadeh, J. Michaelson, A. Sattar, T. Tsunoda, A. Sharma, "PSSM-Suc: accurately predicting succinylation using position specific scoring matrix into bigram for feature extraction," Journal of Theoretical Biology, vol. 425, pp. 97-102, DOI: 10.1016/j.jtbi.2017.05.005, 2017.
[18] Y. López, A. Dehzangi, S. P. Lal, G. Taherzadeh, J. Michaelson, A. Sattar, T. Tsunoda, A. Sharma, "SucStruct: prediction of succinylated lysine residues by using structural properties of amino acids," Analytical Biochemistry, vol. 527, pp. 24-32, DOI: 10.1016/j.ab.2017.03.021, 2017.
[19] J. Jia, Z. Liu, X. Xiao, B. Liu, K. C. Chou, "pSuc-Lys: predict lysine succinylation sites in proteins with PseAAC and ensemble random forest approach," Journal of Theoretical Biology, vol. 394, pp. 223-230, DOI: 10.1016/j.jtbi.2016.01.020, 2016.
[20] M. M. Hasan, M. S. Khatun, M. N. H. Mollah, C. Yong, D. Guo, "A systematic identification of species-specific protein succinylation sites using joint element features information," International Journal of Nanomedicine, vol. Volume 12, pp. 6303-6315, DOI: 10.2147/IJN.S140875, 2017.
[21] Q. Ning, X. Zhao, L. Bao, Z. Ma, X. Zhao, "Detecting succinylation sites from protein sequences using ensemble support vector machine," BMC Bioinformatics, vol. 19 no. 1,DOI: 10.1186/s12859-018-2249-4, 2018.
[22] A. G. de Brevern, M. M. Hasan, H. Kurata, "GPSuc: Global Prediction of Generic and Species-specific Succinylation Sites by aggregating multiple sequence features," PLoS One, vol. 13 no. 10, 2018.
[23] A. Dehzangi, Y. López, S. P. Lal, G. Taherzadeh, A. Sattar, T. Tsunoda, A. Sharma, "Improving succinylation prediction accuracy by incorporating the secondary structure via helix, strand and coil, and evolutionary information from profile bigrams," PLoS One, vol. 13 no. 2, article e0191900,DOI: 10.1371/journal.pone.0191900, 2018.
[24] Y. López, A. Sharma, A. Dehzangi, S. P. Lal, G. Taherzadeh, A. Sattar, T. Tsunoda, "Success: evolutionary and structural properties of amino acids prove effective for succinylation site prediction," BMC Genomics, vol. 19 no. S1,DOI: 10.1186/s12864-017-4336-8, 2018.
[25] H.-J. Kao, V.-N. Nguyen, K.-Y. Huang, W.-C. Chang, T.-Y. Lee, "SuccSite: incorporating amino acid composition and informative k-spaced amino acid pairs to identify protein succinylation sites," Genomics, Proteomics & Bioinformatics, vol. 18 no. 2, pp. 208-219, DOI: 10.1016/j.gpb.2018.10.010, 2020.
[26] K.-Y. Huang, J. B.-K. Hsu, T.-Y. Lee, "Characterization and identification of lysine succinylation sites based on deep learning method," Scientific Reports, vol. 9 no. 1, article 16175,DOI: 10.1038/s41598-019-52552-4, 2019.
[27] L. Zhang, M. Liu, X. Qin, G. Liu, H. Ding, "Succinylation Site Prediction Based on Protein Sequences Using the IFS- LightGBM (BO) Model," Computational and Mathematical Methods in Medicine, vol. 2020,DOI: 10.1155/2020/8858489, 2020.
[28] M. M. Hasan, M. S. Khatun, H. Kurata, "Large-scale assessment of bioinformatics tools for lysine succinylation sites," Cell, vol. 8 no. 2,DOI: 10.3390/cells8020095, 2019.
[29] N. Thapa, M. Chaudhari, S. McManus, K. Roy, R. H. Newman, H. Saigo, D. B. Kc, "DeepSuccinylSite: a deep learning based approach for protein succinylation site prediction," BMC Bioinformatics, vol. 21 no. S3,DOI: 10.1186/s12859-020-3342-z, 2020.
[30] Z. Liu, J. Cao, X. Gao, Y. Zhou, L. Wen, X. Yang, X. Yao, J. Ren, Y. Xue, "CPLA 1.0: an integrated database of protein lysine acetylation," Nucleic Acids Research, vol. 39, pp. D1029-D1034, DOI: 10.1093/nar/gkq939, 2011.
[31] Z. Liu, Y. Wang, T. Gao, Z. Pan, H. Cheng, Q. Yang, Z. Cheng, A. Guo, J. Ren, Y. Xue, "CPLM: a database of protein lysine modifications," Nucleic Acids Research, vol. 42 no. D1, pp. D531-D536, DOI: 10.1093/nar/gkt1093, 2014.
[32] H. Xu, J. Zhou, S. Lin, W. Deng, Y. Zhang, Y. Xue, "PLMD: an updated data resource of protein lysine modifications," Journal of Genetics and Genomics, vol. 44 no. 5, pp. 243-250, DOI: 10.1016/j.jgg.2017.03.007, 2017.
[33] Y. Huang, B. Niu, Y. Gao, L. Fu, W. Li, "CD-HIT suite: a web server for clustering and comparing biological sequences," Bioinformatics, vol. 26 no. 5, pp. 680-682, DOI: 10.1093/bioinformatics/btq003, 2010.
[34] W. Li, A. Godzik, "Cd-hit: a fast program for clustering and comparing large sets of protein or nucleotide sequences," Bioinformatics, vol. 22 no. 13, pp. 1658-1659, DOI: 10.1093/bioinformatics/btl158, 2006.
[35] G. Huang, L. Lu, K. Feng, J. Zhao, Y. Zhang, Y. Xu, N. Zhang, B.-Q. Li, W. Huang, Y.-D. Cai, "Prediction of S-nitrosylation modification sites based on kernel sparse representation classification and mRMR algorithm," BioMed Research International, vol. 2014,DOI: 10.1155/2014/438341, 2014.
[36] Q. Xiang, K. Feng, B. Liao, Y. Liu, G. Huang, "Prediction of lysine malonylation sites based on pseudo amino acid," Combinatorial Chemistry & High Throughput Screening, vol. 20 no. 7, pp. 622-628, DOI: 10.2174/1386207320666170314102647, 2017.
[37] T. Mikolov, I. Sutskever, K. Chen, G. S. Corrado, J. Dean, "Distributed representations of words and phrases and their compositionality," Advances in neural information processing systems, pp. 3111-3119, 2013.
[38] T. Mikolov, K. Chen, G. Corrado, D. JJapa, "Efficient estimation of word representations in vector space," 2013. https://arxiv.org/abs/1301.3781
[39] Y. LeCun, B. Boser, J. S. Denker, D. Henderson, R. E. Howard, W. Hubbard, L. D. Jackel, "Backpropagation applied to handwritten zip code recognition," Neural Computation, vol. 1 no. 4, pp. 541-551, DOI: 10.1162/neco.1989.1.4.541, 1989.
[40] Y. Lecun, L. Bottou, Y. Bengio, P. Haffner, "Gradient-based learning applied to document recognition," Proceedings of the IEEE, vol. 86 no. 11, pp. 2278-2324, DOI: 10.1109/5.726791, 1998.
[41] A. Krizhevsky, I. Sutskever, G. E. Hinton, "Imagenet classification with deep convolutional neural networks," Communications of the ACM, vol. 60 no. 6, pp. 84-90, DOI: 10.1145/3065386, 2017.
[42] K. Simonyan, A. Zisserman, "Very deep convolutional networks for large-scale image recognition," 2014. https://arxiv.org/abs/1409.1556
[43] C. Szegedy, W. Liu, Y. Jia, P. Sermanet, S. Reed, D. Anguelov, D. Erhan, V. Vanhoucke, A. Rabinovich, "Going deeper with convolutions," Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), .
[44] K. He, X. Zhang, S. Ren, J. Sun, "Deep residual learning for image recognition," IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 778-779, 2016.
[45] B. A. Pearlmutter, "Learning state space trajectories in recurrent neural networks," Neural Computation, vol. 1 no. 2, pp. 263-269, DOI: 10.1162/neco.1989.1.2.263, 1989.
[46] C. L. Giles, G. M. Kuhn, R. J. Williams, "Dynamic recurrent neural networks: theory and applications," IEEE Transactions on Neural Networks, vol. 5 no. 2, pp. 153-156, DOI: 10.1109/TNN.1994.8753425, 1994.
[47] S. Hochreiter, J. Schmidhuber, "Long short-term memory," Neural Computation, vol. 9 no. 8, pp. 1735-1780, DOI: 10.1162/neco.1997.9.8.1735, 1997.
[48] M. Schuster, K. K. Paliwal, "Bidirectional recurrent neural networks," IEEE Transactions on Signal Processing, vol. 45 no. 11, pp. 2673-2681, DOI: 10.1109/78.650093, 1997.
[49] G. E. Hinton, N. Srivastava, A. Krizhevsky, I. Sutskever, R. R. Salakhutdinov, "Improving neural networks by preventing co-adaptation of feature detectors," 2012. https://arxiv.org/abs/1207.0580
[50] N. Srivastava, "Improving neural networks with dropout," University of Toronto, vol. 182 no. 566, 2013.
[51] X. Bouthillier, K. Konda, P. Vincent, R. Memisevic, "Dropout as data augmentation," 2015. https://arxiv.org/abs/1506.08700
[52] P. Baldi, P. Sadowski, "The dropout learning algorithm," Artificial Intelligence, vol. 210, pp. 78-122, DOI: 10.1016/j.artint.2014.02.004, 2014.
[53] N. Srivastava, G. Hinton, A. Krizhevsky, I. Sutskever, R. Salakhutdinov, "Dropout: a simple way to prevent neural networks from overfitting," Journal of Machine Learning Research, vol. 15 no. 1, pp. 1929-1958, 2014.
[54] H. Mi, A. Muruganujan, D. Ebert, X. Huang, P. D. Thomas, "PANTHER version 14: more genomes, a new PANTHER GO-slim and improvements in enrichment analysis tools," Nucleic Acids Research, vol. 47 no. D1, pp. D419-D426, 2018.
[55] H. Mi, A. Muruganujan, X. Huang, D. Ebert, C. Mills, X. Guo, P. D. Thomas, "Protocol update for large-scale genome and gene function analysis with the PANTHER classification system (v.14.0)," Nature Protocols, vol. 14 no. 3, pp. 703-721, DOI: 10.1038/s41596-019-0128-8, 2019.
[56] D. W. Huang, B. T. Sherman, R. A. Lempicki, "Bioinformatics enrichment tools: paths toward the comprehensive functional analysis of large gene lists," Nucleic Acids Research, vol. 37 no. 1, 2009.
[57] D. W. Huang, B. T. Sherman, R. A. Lempicki, "Systematic and integrative analysis of large gene lists using DAVID bioinformatics resources," Nature Protocols, vol. 4 no. 1, pp. 44-57, 2009.
You have requested "on-the-fly" machine translation of selected content from our databases. This functionality is provided solely for your convenience and is in no way intended to replace human translation. Show full disclaimer
Neither ProQuest nor its licensors make any representations or warranties with respect to the translations. The translations are automatically generated "AS IS" and "AS AVAILABLE" and are not retained in our systems. PROQUEST AND ITS LICENSORS SPECIFICALLY DISCLAIM ANY AND ALL EXPRESS OR IMPLIED WARRANTIES, INCLUDING WITHOUT LIMITATION, ANY WARRANTIES FOR AVAILABILITY, ACCURACY, TIMELINESS, COMPLETENESS, NON-INFRINGMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Your use of the translations is subject to all use restrictions contained in your Electronic Products License Agreement and by using the translation functionality you agree to forgo any and all claims against ProQuest or its licensors for your use of the translation functionality and any output derived there from. Hide full disclaimer
Copyright © 2021 Guohua Huang et al. This is an open access article distributed under the Creative Commons Attribution License (the “License”), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited. Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License. https://creativecommons.org/licenses/by/4.0/
Abstract
Lysine succinylation is a typical protein post-translational modification and plays a crucial role of regulation in the cellular process. Identifying succinylation sites is fundamental to explore its functions. Although many computational methods were developed to deal with this challenge, few considered semantic relationship between residues. We combined long short-term memory (LSTM) and convolutional neural network (CNN) into a deep learning method for predicting succinylation site. The proposed method obtained a Matthews correlation coefficient of 0.2508 on the independent test, outperforming state of the art methods. We also performed the enrichment analysis of succinylation proteins. The results showed that functions of succinylation were conserved across species but differed to a certain extent with species. On basis of the proposed method, we developed a user-friendly web server for predicting succinylation sites.
You have requested "on-the-fly" machine translation of selected content from our databases. This functionality is provided solely for your convenience and is in no way intended to replace human translation. Show full disclaimer
Neither ProQuest nor its licensors make any representations or warranties with respect to the translations. The translations are automatically generated "AS IS" and "AS AVAILABLE" and are not retained in our systems. PROQUEST AND ITS LICENSORS SPECIFICALLY DISCLAIM ANY AND ALL EXPRESS OR IMPLIED WARRANTIES, INCLUDING WITHOUT LIMITATION, ANY WARRANTIES FOR AVAILABILITY, ACCURACY, TIMELINESS, COMPLETENESS, NON-INFRINGMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Your use of the translations is subject to all use restrictions contained in your Electronic Products License Agreement and by using the translation functionality you agree to forgo any and all claims against ProQuest or its licensors for your use of the translation functionality and any output derived there from. Hide full disclaimer
Details
; Shen, Qingfeng 1 ; Zhang, Guiyang 1 ; Wang, Pan 1 ; Zu-Guo, Yu 2 1 School of Information Engineering, Shaoyang University, Shaoyang 42200, China
2 Key Laboratory of Intelligent Computing and Information Processing of Ministry of Education and Hunan Key Laboratory for Computation and Simulation in Science and Engineering, Xiangtan University, Xiangtan, Hunan 411105, China





