 Research
 Open Access
 Published:
Predicting betaturns in proteins using support vector machines with fractional polynomials
Proteome Science volume 11, Article number: S5 (2013)
Abstract
Background
βturns are secondary structure type that have essential role in molecular recognition, protein folding, and stability. They are found to be the most common type of nonrepetitive structures since 25% of amino acids in protein structures are situated on them. Their prediction is considered to be one of the crucial problems in bioinformatics and molecular biology, which can provide valuable insights and inputs for the fold recognition and drug design.
Results
We propose an approach that combines support vector machines (SVMs) and logistic regression (LR) in a hybrid prediction method, which we call (HSVMLR) to predict βturns in proteins. Fractional polynomials are used for LR modeling. We utilize position specific scoring matrices (PSSMs) and predicted secondary structure (PSS) as features. Our simulation studies show that HSVMLR achieves Qtotal of 82.87%, 82.84%, and 82.32% on the BT426, BT547, and BT823 datasets respectively. These values are the highest among other βturns prediction methods that are based on PSSMs and secondary structure information. HSVMLR also achieves favorable performance in predicting βturns as measured by the Matthew's correlation coefficient (MCC) on these datasets. Furthermore, HSVMLR shows good performance when considering shape strings as additional features.
Conclusions
In this paper, we present a comprehensive approach for βturns prediction. Experiments show that our proposed approach achieves better performance compared to other competing prediction methods.
Background
Secondary structure of proteins consists of basic elements; these elements are αhelices, βsheets, random coils, and turns. αhelices and βsheets are considered as regular secondary structure elements while the residues that correspond to turns structures do not form regular secondary structure elements. In turns structures the Cαatoms of two residues are separated by one to five peptide bonds and the distance between these Cαatoms is less than 7A°. The number of peptide bonds that separate the two end residues determines the specific turn type. In αturns and βturns, the two end residues are separated by four and three peptide bonds respectively. In γturns, δturns, and πturns, the two end residues are separated by two, one, and five peptide bonds respectively. The most common types of turns structure that exist in protein are βturns structure. They represent approximately 25% of the secondary structure of the proteins sequences. βturns can reverse the direction of a protein chain therefore they are considered as orienting structure [1]. They also have significant effects in protein folding, because they have the ability to bring together and allow the interactions between the regular secondary structure elements. βturns are not only important in protein folding but are also implicated in the biological activities of peptides as the bioactive structures that interact with other molecules such as receptors, enzymes and antibodies [2]. They are also important in the design of various peptidomimetics for many diseases [3]. Therefore, the prediction of βturns is one of the important problems in molecular biology, which can provide valuable insights and inputs for the fold recognition and drug design.
There are different methods designed for βturns prediction. These methods can be divided into statistical methods and machine learning methods. The statistical methods that are used in βturns prediction include ChouFasman method [4], Thornton's algorithm [5], GORBTURN [6], 14 & 23 correlation model [7], sequence couple model [8], and COUDES method [9]. All of these statistical methods use the sequence as input except for COUDES, which is based on propensities and multiple alignments. COUDES also utilizes secondary structure predicted by PSIPRED [10], SSPRO2 [11], and PROF [12]. The machine learning methods include BTPRED [13], BetaTpred2 [14], MOLEBRNN [15] and NetTurnP [1], which are based on artificial neural networks (ANNs), Kim's method based on knearest neighbor (KNN) [16], as well as support vector machines (SVMs) based methods, which recently have become popular in the field of βturns prediction. These SVMs based methods include BTSVM [17], Zhang and colleagues' method [18], Zheng and Kurgan's method [2], Hu and Li's method [19], the method of Liu et al. [20], DEBT [21], and the method of Tang et al. [22]. In BTBRED, secondary structure predictions are utilized with two layered network architecture. BetaTpred2 enhances the performance of βturns prediction by using secondary structure prediction and evolutionary information in form of position specific scoring matrices (PSSMs) as input to the neural networks. MOLEBRNN uses PSSMs as input to a bidirectional Elmantype recurrent neural network. NetTurnP uses evolutionary information and predicted protein sequence features as input to two ANN layers whereas the first layer is trained to predict whether or not an amino acid is located in a βturn. Kim's method encodes protein sequence using a window of up to 9 residues to be used as input to a KNN based method, which is combined with a filter that uses secondary structure predicted with PSIPRED for the central residue. In BTSVM, position specific frequent matrices (PSFMs) and PSSMs, both calculated with PSIBLAST [23], are applied to encode input for SVM classifier. Zhang and colleagues' method is another SVM method that uses PSSMs over a 7residue window and the secondary structure of the central residue predicted by PSIPRED as an input. In Zheng and Kurgan's method a SVM is utilized to predict βturns using window based information extracted from four predicted secondary structures (PSSs) with a selected set of PSSMs as input to the SVM. The SVM based method developed by Hu and Li combines the increment of diversity, position conservation scoring function, and secondary structure predicted with PSIPRED to compute the inputs for prediction of βturns and γturns. Liu et al. combine SVM with PSS information obtained by using ESSpred, a secondary protein structure prediction method. DEBT predicts βturns and their types using information from multiple sequence alignments, PSSs, and predicted dihedral angles. Tang et al. considered another type of onedimensional string of symbols representing the clustered region of ϕ, ψ torsion pairs called shape strings as new features. In [24] we utilized the idea of undersampling to create several balanced datasets. These balanced sets were used to train several SVMs classifiers independently. The SVMs were aggregated using a linear logistic regression model.
In this paper, we propose a new approach called HSVMLR (Hybrid approach of SVMs and Logistic Regression (LR)) for predicting βturns. Our proposed approach incorporates the idea of clustering by partitioning the nonβturn class into three subsets using kmeans clustering algorithm. Each subset is merged with the positive class (βturn) to form a sub training set. These sub training sets are used to train localized SVMs classifiers independently. LR model modeled using fractional polynomials, is used to aggregate the localized SVMs to make a collective decision. The merit of using LR to aggregate the localized SVMs is that it will enable us to take advantages of the statistical modeling theory to find the optimal weights for each local SVM [24]. Also LR has the advantages of being widely studied [25], and in the recent years there are many algorithms have been designed to improve its performance. These algorithms include iteratively reweighted least squares (IRLS) algorithm, which is a special case of fisher's scoring method [26, 27].
Methods
Support vector machine (SVM)
The SVM is a stateoftheart supervised learning model with associated learning algorithm for analyzing and classifying data. It transfers the data from low dimensional space to high or infinite dimensional space and then construct a hyperplane or hyperplanes in this higher dimensional space to classify the transformed data. Normally the training data are represented as points in a vector space. The hyperplane with the largest distance to the nearest training data point is considered to be the good separator. Given a training set {x _{ i }, y _{ i }}_{ i = 1, ..., l }, where x _{ i } is a vector of features, and y _{ i } ∈ {1, 1}. SVM solves the following primal problem.
subject to
where w is the normal vector to the hyperplane, b is the offset from the origin, and C is the error penalty parameter. The kernel function, which maps the input space into a higherdimensional space, can be applied to create SVM classifier for nonlinear problem. The kernel functions that can be used for SVM include polynomial kernel function, radial basis (also known as Gaussian kernel function), and sigmoid kernel function.
Logistic regression (LR)
LR is a type of regression analysis used for predicting the outcome of a variable that can take on a limited number of classes. A detailed description of logistic regression can be found in [25]. In brief, given input vectors x _{ i } ∈ R^{n} and output values y _{ i } ∈ {0, 1}, logistic regression can be fitted using the following likelihood to predict the probability of the output. This probability will be p if y _{ i } = 1, or 1  p if y _{ i } = 0.
However, it is easier mathematically to work with log of equation. The loglikelihood, where the log will turn products into sums, can be defined as follows:
The value of θ that maximizes L(θ) is called the maximum likelihood estimate and it is denoted as $\hat{\theta}$. For binary outputs, the loss function or the deviance (DEV) is the negative loglikelihood and is given by the following formula.
Minimizing the deviance given in the above equation is equivalent to maximizing the loglikelihood.
Datasets
The dataset BT426, which contains 426 nonhomologous protein chains, is used to evaluate our HSVMLR prediction method. This dataset was developed by Guruprasad and Rajkumar [28]. We obtained it from Raghava Group's website http://www.imtech.res.in/raghava/bteval/dataset.html. The structure of protein chains in BT426 dataset is determined by Xray crystallography at two resolution or better. In each chain there is at least one betaturns structure. 24.9% of all amino acids in BT426 have been assigned to be having βturns structure. Several recent betaturns prediction methods use it as a golden set of amino acid sequences to evaluate their performances. We therefore used it to evaluate our methods and to make direct comparisons with the other prediction methods. Besides BT426, we used the dataset of 547 protein sequence (BT547), and the dataset of 823 protein sequence (BT823) to evaluate our approach. These datasets were constructed for training and testing COUDES [9].
Features
PSSMs
It has been shown that PSSMs contributed significantly to the accuracy of βturns prediction [1, 2]. They are in the form of M*20, where M represents the sequence length. The PSSMs are generated using three rounds of the iterative PSIBLAST program [23] against National Center for Biotechnology Information (NCBI) nonredundant (nr) sequence database with the default parameters. The PSSMs values are scaled to values between 0 and 1 using the following function.
where x is the PSSM's element that stands for the likelihood of the particular residue substitution at that position.
Predicted secondary structure (PSS)
PROTEUS [29] is used to predict the secondary structure features. The motivation to use PROTEUS comes from the work of Tang et al. [22], which concludes that the predictions when using PROTEUS and PSSMs were better than when using PHD [30], JPRED [31], PROTEUS, and PSSMs together. The secondary structure features are predicted as three structure states: helix (H), strand (E) and coil (C). These three structure states are encoded as 1 0 0 for helix, 0 1 0 for strand, and 0 0 1 for coil.
Predicted shape strings
Tang et al. [22] predicted shape strings from a predictor constructed based on structural alignment approach. Shape strings were represented by eight states, i.e. S, R, U, V, K, A, T and G. They used a sliding window of 8 amino acids on PSSMs, PSS and shape strings features. We also added shape strings to our PSSMs and PSS features. The shape strings were predicted using the protein shape string and its profile prediction server (DSP) [32]. Besides the eight states DSP defines shape N where the ϕ and ψ angles are undefined, or no structure determination for parts of the sequence. The shape strings features are encoded as (1 0 0 0 0 0 0 0 0) for S, (0 1 0 0 0 0 0 0 0) for R, ..., and (0 0 0 0 0 0 0 0 1) for N.
The proposed approach
The entire framework of our proposed approach is shown in Figure 1. Three SVM classifiers are constructed using inputs from three clustered model. Then these three SVMs classifiers are integrated with logistic regression model. Statistical model selection based on fractional polynomials is used to take advantage of each classifier such that the final global classifier could have a better performance.
A sliding window of size seven residues is used over the matrix that consists of the features. The prediction is made for the central residue. This window size is selected in accordance with Shepherd et al. [13] who found that the optimal prediction for βturns is achieved using window size of seven or nine.
Clustered model
Since βturns account for approximately 25% of the globular protein residues, the ratio of βturns to nonβturns is 1:3. Thus, the training sets used for βturns prediction are imbalanced sets. In our trail experiments, we found that if the nonβturns set is divided into a three subsets by a suitable clustering algorithm, each nonβturns subset with the whole βturns set will form approximately balanced training set. This balanced training set is more likely to be separable in the feature space. That is because the distribution of the nonβturns samples in a subset is centralized and compacted. In other words, the βturns set can be easily separated from each nonβturns cluster by a different hyperplane. That means good performance would be expected when constructing localized SVMs using each nonβturns cluster against the βturns. But, each of these SVMs alone is certainly not a good global classifier. It proposes that it is possible to construct a better classifier than the SVM trained with the whole data by combining these SVMs effectively. Particularly, a localized SVM classifier can be constructed for each sub training set, this way the localized SVMs will not be affected by the heterogeneity of the whole training set. To outperform the SVM that is trained with the whole data, we need to combine these localized SVMs effectively into global one without neglecting their local advantages. Majority voting is one of the methods that are used to combine several classifiers, but its main problem is that it will not give weight to each classifier. LR model can integrate the localized SVMs classifiers, and it allows us to take advantages of the statistical modeling theory to find the optimal weights for each local classifier. The motivation to use this clustered model comes from the work of Yi Chang [33]. In his work, Yi Chang used localized linear SVMs classifier for a data in the feature space defined by a chosen kernel.
At the very beginning, the whole negative examples are divided into three clusters by a kmeans clustering algorithm using original variables. The distribution of those three clusters is shown in Figure 2. We merged the whole positive examples with each cluster to form three subtraining sets. These subtraining sets are used to build three SVMs models. The three SVMs will not be used directly in the prediction, but they will be used as variable generators. During training and prediction stages, these models are unchanged and all the samples enter all of the three models. The signed distance for each example to the separating hyperplanes of the three models is computed and stored in a vector d of dimension (N * 3), where N is the number of the instances. The vector d will be used as a new feature vector for a LR model, which will weigh the response of the three models and then calculates the prediction probability.
LR model selection
The components of the LR predictive model are obviously variables, which should be selected carefully so that the model makes accurate prediction, but without overfitting the data. There are two competing goals in model selection. (1) It should be complex to fit the data well. (2) It should be simple to interpret. To select our LR model, we first looked at the correlation in the estimated coefficient. If two variables are highly correlated, we do not need both of them in the model. The univariate analysis was used to identify the important variables, in which the LR models with one variable at a time were fitted, and then the fits were analyzed. In particular, we looked at the estimated coefficients, their standard errors and the likelihood ratio test for the significance of the coefficients. Then we fitted our LR using the variables selected in the univariate analysis according to the following procedure:

We verified the importance of each variable in the LR model using Wald statistics.

We compared the coefficients of the each variable with the coefficient from the model containing only that variable.

Any variable that did not appear to be important was eliminated, and a new model was fitted. The new model was checked whether it is significantly different from the old model. If it is, then the deleted variable is important.

The process of deleting, refitting and verifying was repeated until it appears that all the important variables were included in the model.

We tried to fit a linear LR model to the data but the prediction error is found to be very large, so we considered power transformation using fractional polynomials.

A list of possible interactions between each pairs of variable was created, these interactions terms were added one at a time, in the model containing all the main effects and assess its significance using the likelihood ratio test. The significant interactions were added to the main effect model and its fit was evaluated using Wald tests and LR test for the interaction terms, and any nonsignificant interaction was dropped.
Fractional polynomials
The final outcome variable is the βturn/nonturn response. In our hybrid model, this variable depends on the outcome of the three SVMs classifiers in a logistic regression model. The outcome of the three SVMs classifiers is represented by the vector d = (d _{1}, d _{2}, d _{3}). The natural starting point, the straight line model b _{0} + b _{1} d _{1} + b _{2} d _{2} + b _{3} d _{3} or b _{0} + dB in matrix form, where B is the vector of parameters, is first tested whether it is adequate. To improve the fit, we investigated other models. We looked for nonlinearity by fitting a first order fractional polynomial to the data. The best power transformation ${d}_{i}^{p}$ was found, with the power p chosen from candidates 2, 1, 0.5, 0, 0.5, 1, 2, 3, where ${d}_{i}^{0}$ denotes log(d _{ i }). The set includes the straight line (i.e. no transformation) p = 1. The variables d _{ i } contain nonpositive values, thus we transformed its values to values > 0, which will enable the use of logarithms and negative powers transformation. Including more powers usually offers only slight improvement in the model fit. In particular, there is a problem with including large negative powers, such as 3, that individual extreme observations will influence the fit too much [34]. The firstdegree fractional polynomial provides unsatisfactory fit to our data, so we considered seconddegree fractional polynomial. We used the closed test procedure, which first determine the bestfitting second degree polynomial by choosing the powers transformation p and q from the aforementioned set. For mathematical limit, when p = q for the variable d _{ i } in the model then the terms of the variable will be written in the form ${b}_{j}{d}_{i}^{p}+{b}_{k}{d}_{i}^{p}log\left({d}_{i}\right)$. The best fit among the combinations of such powers is defined as that which maximizes the likelihood or equivalently that which minimizes the deviance [35]. The MFP package, which is a collection of R [36] functions targeted at the use of fractional polynomials for modeling the influence of continuous variables on the outcome in regression models is used in this research to find the best fit among the combinations of the powers p and q. MFP uses a sequential and a closed testing selection procedures for a single continuous variable. Using the BT426 dataset, our final model is selected after two cycles. The results of the model selection are shown in Table 1. The bestfit fractional polynomials (fractional polynomials with the lowest deviance) for SVM model1, SVM model2, and SVM model3 are underlined.
Training and testing
We used LIBSVM package [37] to train and build the SVMs prediction models. The radial basis kernel function was used to transfer the data from a low dimension space to a higherdimensional space nonlinearly for all the SVMs. The default grid search approach was used to find the optimal values for the LIBSVM's parameters C and gamma. The leaveoneout crossvalidation test, in which different datasets for training and testing are used to evaluate a prediction method, is an accurate test method compared with independent dataset test and subdataset test [38]. When using this test, one protein out of N proteins is removed to represent the testing set and the remaining N1 proteins are combined together to represent the training set that will be used for training the prediction method. This process is then repeated N times by removing one protein in each time. In βturns prediction, applying this process exactly is time consuming. Thus, most of the stateoftheart βturns prediction methods use sevenfold cross validation to assess their prediction performances [39]. Therefore, we used sevenfold cross validation to assess the performance of our HSVMLR method. We first started by dividing the dataset into seven subsets that contain equal numbers of proteins. In each set the βturns account for approximately 25% of the protein residues, in other words each set contains the naturallyaccruing proportion of betaturns. We removed one set to represent the testing set and the other sets were merged together in one training set, which is used to train HSVMLR. This process was repeated seven times in order to have a different set for testing each time. We take the average of the results from the seven testing sets to represent the final prediction result.
Performance measures
The quality of prediction is evaluated using four measures, the prediction accuracy, Qpredicted, Qobserved, and MCC. These measures are the most frequently used measures to evaluate the βturns prediction methods. They are calculated using the four values (i) true positive (TP), which is the number of the residues that are correctly classified as βturns, (ii) true negative (TN), which is the number of the residues that are correctly classified as nonβturns, (iii) false positive (FP), which is the number of residues that have nonβturns structure and incorrectly classified as having βturns structure, and (iv) false negative (FN), which is the number of residues that have βturns structure and incorrectly classified as having nonβturns structure.
The prediction accuracy (also known as Qtotal) refers to the percentage of correctly classified residues and is calculated as follows:
Qpredicted (also known as the predicted positive value (PPV) or the probability of correct prediction) refers to the percentage of the residues that are correctly predicted as βturns among the predicted ones and is calculated as follows:
Qobserved (also known as sensitivity or coverage) refers to the percentage of the residues that are correctly predicted to have βturns structure among those observed as having βturns structure. In other words, it represents the fraction of the total positive samples that are correctly predicted and it is calculated as follows:
Because of the imbalanced dataset (25% βturns), Qtotal by itself is a poor measure. In other words, one can achieve a Qtotal of 75% (baseline accuracy) by predicting all the residues to be nonβturns. Therefore, Matthew's correlation coefficient (MCC) [40] is an important, robust and reliable performance measure. The MCC can be obtained using the following formula:
Normally, the value of MCC is greater than or equal to 1 and less than or equal to 1. If the value of MCC is close to 1 then there is a perfect positive correlation, if it is close to 1 then there is a perfect negative correlation, and a value close to 0 indicates no correlation.
The receiver operating characteristic (ROC) curve is adopted in this paper as a threshold independent measure. The ROC curve provides the effectiveness of βturns prediction method. The area under the ROC curve (AUC) is an important index that reflects the prediction reliability. A good classifier has an area close to 1, while a random classifier has an area of 0.5.
Results and discussion
The methods that are applied on βturns prediction use different PSSMs and PSS organizations. Some researchers use a sliding window on the PSSMs and then add the PSS e.g. [18]. Other researchers use a sliding window on both PSSMs and PSS e.g. [20]. Both ways are tested in our proposed method and the results for the BT426 dataset are shown in Table 2.
From the results we found that the performance of HSVMLR using a sliding window on both PSSMs and PSS is by far better than using a sliding window on PSSMs only and then add the PSS for the central amino acid. Figure 3 shows the ROC curves for βturns prediction using a sliding window on PSSMs only and a sliding window on both PSSMs and PSS. The AUC highlights the effect of using a sliding window on both PSSMs and PSS. The AUC value using a sliding window on both PSSMs and PSS is 0.89, 0.03 higher than using a sliding window on the PSSMs only.
Table 3 shows the comparison between HSVMLR and other existing βturns prediction methods based on sevenfold cross validation on the BT426 dataset. HSVMLR achieves prediction accuracy or Qtotal = 82.87%, Qpredicted= 64.83%, Qobserved = 70.66%, and MCC = 0.56. The Qtotal of HSVMLR is the highest among the existing methods that use PSSMs and PSS as features; i.e. Zheng and Kurgan's method and the method of Liu et al. achieved Qtotal of 80.9. The difference in Qtotal between HSVMLR and these methods is 1.97%. We emphasize that this difference is relatively large when considering that the baseline accuracy equals to 75%, which could be obtained by merely regarding all residues as nonβturns. i.e., HSVMLR provides 7.87/25 = 31.5% error rate reduction, while Zheng and Kurgan's method and the method of Liu et al. provide 5.9/25 = 24% error rate reduction, and Hu and Li's method provides 4.8/25 = 19% error rate reduction.
HSVMLR shows high MCC 0.56 compared to NetTurnP 0.50, Zheng and Kurgan's method 0.47, and the method of Liu et al. 0.44. Thus, HSVMLR has the highest MCC and Qtotal among the other βturns prediction methods. The MCC value achieved is noteworthy since MCC accounts for both over predictions and under predictions. The Qobserved of HSVMLR is higher by 15.06% than the Qobserved of Zheng and Kurgan's method, by 1.76% than the Qobserved of Hu and Li's method, and by 21.46% than the Qobserved of the method of Liu et al. Higher Qobserved values mean that a large percentage of the observed βurns is correctly predicted. At the same time, the Qpredicted of our method shows that more than 64% of the actual βturns are correctly predicted. We note that the Qpredicted of HSVMLR is 2.13% higher than the Qpredicted of Zheng and Kurgan's method, by 9.23% than the Qpredicted of Hu and Li's method, and by 1.23% higher than the Qpredicted of the method of Liu et al.
Besides BT426 dataset that is used for training and testing HSVMLR, we used two additional datasets, i.e. BT547 and BT823 datasets, to validate its performance. Results obtained based on sevenfold cross validation on these datasets are given in Table 4. The results show that for the BT547 dataset HSVMLR obtains Qtotal = 82.84%, Qpredicted = 63.60%, Qobserved = 68.50%, and MCC = 0.55. The MCC and Qtotal of HSVMLR are the best among the other competing methods that are evaluated on BT547 dataset. We note that the Qpredicted of HSVMLR is 0.7% lower than the Qpredicted of the method of Liu et al., while the Qobserved of HSVMLR is 24% higher than the Qobserved of the method of Liu et al. The increase in the Qobserved value is a tradeoff for the decrease in the Qpredicted value. In spite of this trade off, HSVMLR shows high overall accuracy. For the BT823 dataset HSVMLR obtains Qtotal = 82.32%, Qpredicted = 64.48%, Qobserved = 72.72%, and MCC = 0.56. Also HSVMLR has the highest MCC, Qtotal, Qpredicted, and Qobserved on BT823 datasets. The results also show that HSVMLR shows stable performances on all the three datasets used. Note that we used the same LR model that is used for testing BT426. These results indicate that HSVMLR can better discriminate between βturns and nonβturns.
Including shape strings features
The comparisons between HSVMLR after including the shape strings features and the method of Tang et al. on the BT426, BT547, and BT823 are shown in Table 5. Figure 4 depicts the ROC curves for βturns prediction using HSVMLR before and after adding the shape strings for the BT426 dataset. The AUC value when including the shape strings is 0.923, while the AUC value when using PSSMs and PSS only is 0.886.
Conclusions
In this paper, we proposed an approach that combines SVM and LR to create a hybrid method for βturns prediction. We called this hybrid method HSVMLR. In HSVMLR, we utilized protein profile in the form of PSSMs, and PSS as features. We also considered shape strings as additional features. We divided the nonβturn class into three partitions using kmeans clustering algorithm and then each partition is combined with the βturn class to form approximately balanced subtraining sets. SVM classifier is used for each subtraining set. Using this procedure, the problem of imbalanced class can be overcome, and the SVM computational time can be reduced. LR model selected based on fractional polynomials is used to aggregate the decisions of the SVMs to come up with final βturn or nonβturn decision. Using LR to aggregate the decisions of the SVMs enables us to take advantages of the statistical modeling theory to find the optimal weights for each SVM. HSVMLR achieved MCC of 0.56, and Qtotal of 82.87% on the BT426 dataset when using PSSMs and PSS as features. The MCC and the Qtotal achieved are significantly higher than the best existing methods that predict betaturns using PSSM and PSS. Also HSVMLR obtained the highest MCC and Qtotal on BT547 and BT823 datasets. Furthermore, HSVMLR shows good performance when including shape strings features.
References
 1.
Petersen B, Lundegaard C, Petersen TN: NetTurnPNeural network prediction of betaturns by use of evolutionary information and predicted protein sequence features. PLoS ONE 2010, 5: e15079. 10.1371/journal.pone.0015079
 2.
Zheng C, Kurgan L: Prediction of betaturns at over 80% accuracy based on an ensemble of predicted secondary structures and multiple alignments. BMCBioinformatics 2008, 9: 430. 10.1186/147121059430
 3.
Kee KS, Jois SD: Design of betaturn based therapeutic agents. Curr Pharm Des 2003, 9: 1209–24. 10.2174/1381612033454900
 4.
Chou PY, Fasman G: Conformational parameters for amino acids in helical, β sheet and random coil regions calculated from proteins. Biochemistry 1974, 13: 211–222. 10.1021/bi00699a001
 5.
Wilmot CM, Thornton JM: Analysis and prediction of the different tybes of β turns in proteins. J Mol Biol 1988, 203: 221–232. 10.1016/00222836(88)901039
 6.
Wilmot CM: β Turns and their distortions:a proposed new nomenclature. Protein Eng 1990, 3: 479–493. 10.1093/protein/3.6.479
 7.
Zhang CT, Chou KC, Zhang CT, Chou KC: Prediction of betaturns in proteins by 1–4 & 2–3 correlation model. Biopolymers 1997, 41: 673–702. 10.1002/(SICI)10970282(199705)41:6<673::AIDBIP7>3.0.CO;2N
 8.
Chou KC: Prediction of betaturns. J Peptide Res 1997, 49: 120–144.
 9.
Fuchs PF, Alix AJ: High accuracy prediction of β turns and their types using propensities and multiple alignments. ProteinsStructure Function and Bioinformatics 2005, 59: 828–839. 10.1002/prot.20461
 10.
Jones DT: Protein secondary structure prediction based on positionspecific scoring matrices. J Mol Biol 1999, 292: 195–202. 10.1006/jmbi.1999.3091
 11.
Pollastri G, Przybylski D, Rost B, Baldi P: Improving the prediction of protein secondary structure in three and eight classes using recurrent neural networks and profiles. Proteins 2002, 47: 228–35. 10.1002/prot.10082
 12.
Ouali M, King RD: Cascaded multiple classifiers for secondary structure prediction. Protein Sci 2000, 9: 1162–76. 10.1110/ps.9.6.1162
 13.
Shepherd AJ, Gorse D, Thornton JM: Prediction of the location and type of betaturns in proteins using neural networks. Protein Sci 1999, 8: 1045–1055. 10.1110/ps.8.5.1045
 14.
Kaur H, Raghava GP: Prediction of betaturns in proteins from multiple alignment using neural network. Protein Sci 2003, 12: 627–634. 10.1110/ps.0228903
 15.
Kirschner A, Frishman D: Prediction of betaturns and betaturn types by a novel bidirectional Elmantype recurrent neural network with multiple output layers (MOLECRNN). Gene 2008,422(1–2):22–9. 10.1016/j.gene.2008.06.008
 16.
Kim S: Protein β turn prediction using nearestneighbor method. Bioinformatics 2004, 20: 40–4. 10.1093/bioinformatics/btg368
 17.
Pham TH, Satou K, Ho TB: Prediction and analysis of betaturns in proteins by support vector machine. Genome Informatics 2003, 14: 196–205.
 18.
Zhang Q, Yoon S, Welsh WJ: Improved method for predicting β turn using support vector machine. Bioinformatics 2005, 21: 2370–4. 10.1093/bioinformatics/bti358
 19.
Hu X, Li Q: Using support vector machine to predict betaturns and gammaturns in proteins. J Comput Chem 2008, 29: 1867–1875. 10.1002/jcc.20929
 20.
Liu L, Fang Y, Li M, Wang C: Prediction of betaturn in protein using ESSpred and support vector machine. Protein J 2009, 28: 175–181. 10.1007/s1093000991814
 21.
Kountouris P, Hirst J: Predicting betaturns and their types using predicted backbone dihedral angles and secondary structures. BMC Bioinformatics 2010, 11: 407. 10.1186/1471210511407
 22.
Tang Z, Li T, Liu R, Xiong W, Sun J, Zhu Y, Chen G: Improving the performance of betaturn prediction using predicted shape strings and a twolayer support vector machine model. BMC Bioinformatics 2011, 12: 283. 10.1186/1471210512283
 23.
Altschul SF, Madden TL, Schaffer AA, Zhang J, Zhang Z, Miller W, Lipman DJ: Gapped BLAST and PSIBLAST: a new generation of protein database search programs. Nucleic Acids Res 1997, 25: 3389–3402. 10.1093/nar/25.17.3389
 24.
Elbashir MK, Wang J, Wu FX: A hybrid approach of support vector machines with logistic regression for β turn prediction. BIBMW, IEEE International Conference on Bioinformatics and Biomedicine Workshops 2012, 587–593.
 25.
Hosmer D, Lemeshow S: Applied logistic regression. Wiley 2000.
 26.
Maher M, Theodore B: Robust weighted kernel logistic regression in imbalanced and rare events data. Computational Statistics and Data Analysis 2011, 55: 168–183. 10.1016/j.csda.2010.06.014
 27.
Komarek P, Moore A: Making logistic regression a core data mining tool: a practical investigation of accuracy, speed, and simplicity. In Technical report. Carnegie Mellon University; 2005.
 28.
Guruprasad K, Rajkumar S: Beta and gammaturns in proteins revisited: A new set of amino acid dependent positional preferences and potential. J Biosci 2000,25(2):143–156.
 29.
Montgomerie S, Sundararaj S, Gallin WJ, Wishart DS: Improving the accuracy of protein secondary structure prediction using structural alignment. BMC Bioinformatics 2006, 14: 301.
 30.
Rost B, Sander C: Prediction of protein secondary structure at better than 70% accuracy. J Mol Biol 1993,232(2):584–599. 10.1006/jmbi.1993.1413
 31.
Cole C, Barber JD, Barton GJ: The Jpred 3 secondary structure prediction server. Nucleic Acids Res 2008,36(Web Server issue):W197W201.
 32.
Sun J, Tang S, Xiong W, Cong P, Li T: DSP: a protein shape string and its profile prediction server. Nucleic Acids Res 2012,40(Web server issue):W298. 302
 33.
Chang YI: Boosting SVM classifiers with logistic regression. Technical report, academic Sinica 2003. [http://www3.stat.sinica.edu.tw/library/c_tec_rep/2003–03.pdf]
 34.
Patrick R, Gareth A, Willi S: The use of fractional polynomials to model continuous risk variables in epidemiology. International journal of epidemiology 1999, 28: 964–974. 10.1093/ije/28.5.964
 35.
Royston P, Altman DG: Regression using fractional polynomials of continuous covariates: (parsimonious parametric modelling (with discussion). Appl Stat 1994, 43: 429–467. 10.2307/2986270
 36.
R Development Core Team: R: A language and environment for statistical computing, R Foundation for Statistical Computing. Vienna, Austria; 2008.
 37.
CC C, CJ L, LIBSVM: A library for support vector machines. [http://www.csie.ntu.edu.tw/~cjlin/libsvm]
 38.
Chou K, Zhang C: Prediction of protein structural classes. Critical Reviews in Biochem And Mol Biol 1995, 30: 275–349. 10.3109/10409239509083488
 39.
Elbashir MK, Sheng Y, Wang J, Wu FX, Min Li: Predicting β turns in protein using kernel logistic regression. BioMed Research International 2013., 2013:
 40.
Brunak S, Chauvin Y, Andersen C, Nielsen H: Assessing the accuracy of prediction algorithms: an overview. Bioinformatics 2000, 16: 412–424. 10.1093/bioinformatics/16.5.412
Acknowledgements
This work is supported in part by the National Natural Science Foundation of China under Grant No.61232001, No.61003124, No.61128006, the grant from CityU (Project No. 7002728), the Ph.D. Programs Foundation of Ministry of Education of China No.20090162120073, and the Freedom Explore Program of Central South University No.201012200124.
Declarations
The publication costs for this article were funded by the corresponding author.
This article has been published as part of Proteome Science Volume 11 Supplement 1, 2013: Selected articles from the IEEE International Conference on Bioinformatics and Biomedicine 2012: Proteome Science. The full contents of the supplement are available online at http://www.proteomesci.com/supplements/11/S1.
Author information
Affiliations
Corresponding author
Additional information
Competing interests
The authors declare that they have no competing interests.
Authors' contributions
MKE computed the features, generated the prediction model, performed experimental comparison and drafted the manuscript. JW, FW, LW participated in the design of the study and helped to draft the manuscript. All authors have read and approved the final manuscript.
Rights and permissions
About this article
Cite this article
Elbashir, M.K., Wang, J., Wu, F. et al. Predicting betaturns in proteins using support vector machines with fractional polynomials. Proteome Sci 11, S5 (2013). https://doi.org/10.1186/1477595611S1S5
Published:
Keywords
 Support Vector Machine
 Support Vector Machine Classifier
 Fractional Polynomial
 Torsion Pair
 Shape String