Xiao-Jun Liu
University of Edinburgh
Network
Latest external collaboration on country level. Dive into details by clicking on the dots.
Publication
Featured researches published by Xiao-Jun Liu.
Computational Biology and Chemistry | 2002
Yu-Dong Cai; Xiao-Jun Liu; Xue-biao Xu; Kuo-Chen Chou
In this paper, we apply a new machine learning method which is called support vector machine to approach the prediction of protein structural class. The support vector machine method is performed based on the database derived from SCOP which is based upon domains of known structure and the evolutionary relationships and the principles that govern their 3D structure. As a result, high rates of both self-consistency and jackknife test are obtained. This indicates that the structural class of a protein inconsiderably correlated with its amino and composition, and the support vector machine can be referred as a powerful computational tool for predicting the structural classes of proteins.
BMC Bioinformatics | 2001
Yu-Dong Cai; Xiao-Jun Liu; Xue-biao Xu; Guo-Ping Zhou
BackgroundWe apply a new machine learning method, the so-called Support Vector Machine method, to predict the protein structural class. Support Vector Machine method is performed based on the database derived from SCOP, in which protein domains are classified based on known structures and the evolutionary relationships and the principles that govern their 3-D structure.ResultsHigh rates of both self-consistency and jackknife tests are obtained. The good results indicate that the structural class of a protein is considerably correlated with its amino acid composition.ConclusionsIt is expected that the Support Vector Machine method and the elegant component-coupled method, also named as the covariant discrimination algorithm, if complemented with each other, can provide a powerful computational tool for predicting the structural classes of proteins.
Journal of Computational Chemistry | 2002
Yu-Dong Cai; Xiao-Jun Liu; Xue-biao Xu; Kuo-Chen Chou
Knowledge of the polyprotein cleavage sites by HIV protease will refine our understanding of its specificity, and the information thus acquired is useful for designing specific and efficient HIV protease inhibitors. The pace in searching for the proper inhibitors of HIV protease will be greatly expedited if one can find an accurate, robust, and rapid method for predicting the cleavage sites in proteins by HIV protease. In this article, a Support Vector Machine is applied to predict the cleavability of oligopeptides by proteases with multiple and extended specificity subsites. We selected HIV‐1 protease as the subject of the study. Two hundred ninety‐nine oligopeptides were chosen for the training set, while the other 63 oligopeptides were taken as a test set. Because of its high rate of self‐consistency (299/299=100%), a good result in the jackknife test (286/299=95%) and correct prediction rate (55/63 = 87%), it is expected that the Support Vector Machine method can be referred to as a useful assistant technique for finding effective inhibitors of HIV protease, which is one of the targets in designing potential drugs against AIDS. The principle of the Support Vector Machine method can also be applied to analyzing the specificity of other multisubsite enzymes.
Journal of Cellular Biochemistry | 2002
Yu-Dong Cai; Xiao-Jun Liu; Xue-biao Xu; Kuo-Chen Chou
Support Vector Machine (SVM), which is one class of learning machines, was applied to predict the subcellular location of proteins by incorporating the quasi‐sequence‐order effect (Chou [ 2000 ] Biochem. Biophys. Res. Commun. 278:477–483). In this study, the proteins are classified into the following 12 groups: (1) chloroplast, (2) cytoplasm, (3) cytoskeleton, (4) endoplasmic reticulum, (5) extracellular, (6) Golgi apparatus, (7) lysosome, (8) mitochondria, (9) nucleus, (10) peroxisome, (11) plasma membrane, and (12) vacuole, which account for most organelles and subcellular compartments in an animal or plant cell. Examinations for self‐consistency and jackknife testing of the SVMs method were conducted for three sets consisting of 1,911, 2,044, and 2,191 proteins. The correct rates for self‐consistency and the jackknife test values achieved with these protein sets were 94 and 83% for 1,911 proteins, 92 and 78% for 2,044 proteins, and 89 and 75% for 2,191 proteins, respectively. Furthermore, tests for correct prediction rates were undertaken with three independent testing datasets containing 2,148 proteins, 2,417 proteins, and 2,494 proteins producing values of 84, 77, and 74%, respectively. J. Cell. Biochem. 84: 343–348, 2002.
Peptides | 2002
Yu-Dong Cai; Xiao-Jun Liu; Xue-biao Xu; Kuo-Chen Chou
Support Vector Machines (SVMs) which is one kind of learning machines, was applied to predict the specificity of GalNAc-transferase. The examination for the self-consistency and the jackknife test of the SVMs method were tested for the training dataset (305 oligopeptides), the correct rate of self-consistency and jackknife test reaches 100% and 84.9%, respectively. Furthermore, the prediction of the independent testing dataset (30 oligopeptides) was tested, the rate reaches 76.67%.
Computational Biology and Chemistry | 2002
Yu-Dong Cai; Xiao-Jun Liu; Kuo-Chen Chou
The function of a protein is closely correlated to its subcellular location. Is it possible to utilize a bioinformatics method to predict the protein subcellular location? To explore this problem, proteins are classified into 12 groups (Protein Eng. 12 (1999) 107-118) according to their subcellular location: (1) chloroplast, (2) cytoplasm, (3) cytoskeleton, (4) endoplasmic reticulum, (5) extracellular, (6) Golgi apparatus, (7) lysosome, (8) mitochondria, (9) nucleus, (10) peroxisome, (11) plasma membrane and (12) vacuole. In this paper, the neural network method was proposed to predict the subcellular location of a protein according to its amino acid composition. Results obtained through self-consistency, cross-validation and independent dataset tests are quite high. Accordingly, the present method can serve as a complement tool for the existing prediction methods in this area.
PLOS ONE | 2011
Le-Le Hu; Tao Huang; Xiao-Jun Liu; Yu-Dong Cai
Background Identifying associated phenotypes of proteins is a challenge of the modern genetics since the multifactorial trait often results from contributions of many proteins. Besides the high-through phenotype assays, the computational methods are alternative ways to identify the phenotypes of proteins. Methodology/Principal Findings Here, we proposed a new method for predicting protein phenotypes in yeast based on protein-protein interaction network. Instead of only the most likely phenotype, a series of possible phenotypes for the query protein were generated and ranked acording to the tethering potential score. As a result, the first order prediction accuracy of our method achieved 65.4% evaluated by Jackknife test of 1,267 proteins in budding yeast, much higher than the success rate (15.4%) of a random guess. And the likelihood of the first 3 predicted phenotypes including all the real phenotypes of the proteins was 70.6%. Conclusions/Significance The candidate phenotypes predicted by our method provided useful clues for the further validation. In addition, the method can be easily applied to the prediction of protein associated phenotypes in other organisms.
Journal of Computational Chemistry | 2003
Yu-Dong Cai; Xiao-Jun Liu; Kuo-Chen Chou
The neural network method was applied to the prediction of the content of protein secondary structure elements, including α‐helix, β‐strand, β‐bridge, 310‐helix, π‐helix, H‐bonded turn, bend, and random coil. The “pair‐coupled amino acid composition” originally proposed by K. C. Chou [J Protein Chem 1999, 18, 473] was adopted as the input. Self‐consistency and independent‐dataset tests were used to appraise the performance of the neural network. Results of both tests indicated high performance of the method.
Journal of Biomolecular Structure & Dynamics | 2001
Yu-Dong Cai; Xiao-Jun Liu; Kuo-Chen Chou
Abstract Membrane proteins can be classified among the following five types: (1) type I membrane protein. (2) type II membrane protein. (3) multipass transmembrane proteins. (4) lipid chain- anchored membrane proteins, and (5) GPI-anchored membrane proteins. T. Kohonens self-organization model which is a typical neural network is applied for predicting the type of a given membrane protein based on its amino acid composition. As a result, the high rates of self-consistency (94.80%) and cross-validation (77.76%), and stronger fault-tolerant ability were obtained.
Peptides | 2003
Yu-Dong Cai; Xiao-Jun Liu; Yixue Li; Xue-biao Xu; Kuo-Chen Chou
The support vector machine approach was introduced to predict the β-turns in proteins. The overall self-consistency rate by the re-substitution test for the training or learning dataset reached 100%. Both the training dataset and independent testing dataset were taken from Chou [J. Pept. Res. 49 (1997) 120]. The success prediction rates by the jackknife test for the β-turn subset of 455 tetrapeptides and non-β-turn subset of 3807 tetrapeptides in the training dataset were 58.1 and 98.4%, respectively. The success rates with the independent dataset test for the β-turn subset of 110 tetrapeptides and non-β-turn subset of 30,231 tetrapeptides were 69.1 and 97.3%, respectively. The results obtained from this study support the conclusion that the residue-coupled effect along a tetrapeptide is important for the formation of a β-turn.