Assessment of template-based protein structure predictions in CASP10.
Journal: 2014/September - Proteins: Structure, Function and Genetics
ISSN: 1097-0134
Abstract:
Template-based modeling (TBM) is a major component of the critical assessment of protein structure prediction (CASP). In CASP10, some 41,740 predicted models submitted by 150 predictor groups were assessed as TBM predictions. The accuracy of protein structure prediction was assessed by geometric comparison with experimental X-ray crystal and NMR structures using a composite score that included both global alignment metrics and distance-matrix-based metrics. These included GDT-HA and GDC-all global alignment scores, and the superimposition-independent LDDT distance-matrix-based score. In addition, a superimposition-independent RPF metric, similar to that described previously for comparing protein models against experimental NMR data, was used for comparing predicted protein structure models against experimental protein structures. To score well on all four of these metrics, models must feature accurate predictions of both backbone and side-chain conformations. Performance rankings were determined independently for server and the combined server plus human-curated predictor groups. Final rankings were made using paired head-to-head Student's t-test analysis of raw metric scores among the top 25 performing groups in each category.
Relations:
Content
Citations
(38)
References
(36)
Chemicals
(1)
Processes
(2)
Affiliates
(1)
Similar articles
Articles by the same authors
Discussion board
Proteins 82(0 2): 43-56

Assessment of template based protein structure predictions in CASP10

Introduction

Template based modeling (TBM) is an essential and highly successful approach for protein structure prediction. Recent advances, though generally incremental, provide both server and human curated methods with high reliability for protein structure prediction1. These TBM methods are having a high impact by providing accurate models useful in biological research2. The success of template-based modeling has also been a primary driving force for structural genomics efforts aimed at structural coverage of domain families and biological pathways39. TBM is also a powerful technique for estimating phases in X-ray crystal structure determination by molecular replacement1012. For these reasons, TBM forms an essential component of the Critical Assessment of Protein Structure Prediction (CASP) experiment.

An important activity of the CASP program is the assessment of models and ranking of the performance by various predictor groups. A hallmark of this process is the involvement of independent assessors. A key feature of the CASP model assessment process is that it is done in a “blind” fashion, whereby assessors did not know the identity of each predictor group, which are identified only by a group number. The identities of the competing groups are provided to assessors only after the assessment process is completed.

For CASP10, participants were asked to provide full atomic models of predicted structures, including side-chain atoms. Accurate side-chain structures are often critical for using predicted models in biological applications. Following an emphasis established in CASP813 and CASP91, the TBM assessment process of CASP10 used a combination of metrics which together assess the accuracy of both the backbone and side-chain structures of predicted models.

The assessment of Free Modeling (FM) protein structure predictions is presented in another paper in this same special issue14. In this paper, we describe the TBM assessment results of CASP10. Our analysis largely followed the protocols laid down in CASP813,15 and CASP91,16. In particular, we followed the general procedure using standard measures for TBM assessment outlined by Cozzetto et al.15, and considerations of structure prediction accuracy metrics and statistical comparison tests outlined by Mariani et al1 for the CASP9 TBM assessment.

Methods

Defining Assessment Units

Assessment units were defined based on careful manual analysis of the experimental structures and potential templates, as outlined in the accompanying paper by Taylor et al17. Experimental NMR structures, as well as some X-ray crystal structures, were trimmed back to include only the consistently well-defined regions of the structure using the expanded FindCore algorithm, as described by Snyder et al18 (accompanying paper).

Numerical Automated Structure Quality Assessment Scores

All structure quality assessment scores were computed by the CASP Prediction Center19. GDT and GDC scores were computed for all predictions using the latest version of LGA19,20. The Prediction Center also provided large-scale calculations of LDDT1, RPF, Sphere Grinder19, MolProbity13,21,22, and Prosa23 scores. LDDT scores, which compare the interatomic distance matrices between a predicted model and the experimental structure1,24, were computed with a distance cutoff of 15 Å, which is larger than the 5 Å cutoff used in CASP91. They are referred to throughout this study as LDDT-15 scores. The "--rm=zeroocc" parameter was not used when computing the LDDT score. This means that atoms with zero occupancy were actually included in the LDDT score calculation. As there are not many predictions with zero occupancy atoms, this omission had minimal effect. Sphere Grinder used a 6 Å radius for the sphere, and a 2 Å rmsd cutoff. While these other metrics are described in other publications, the RPF metric for comparing predicted models against experimental structures is described in the next section.

RPF and DP Scores

The RPF method was originally developed as a quality assessment tool for protein NMR structures. The NMR RPF method calculates the Recall, Precision, F-measure and Discriminating Power (DP) scores, by measuring the agreement between all proton-proton distances ≤ 5 Å, and the NMR NOESY and chemical shift data25. Specifically, the DP score measures how well a structural model fits with the NMR data (i.e., NOESY peak list and chemical shift data), normalized by the F-measure score that would be obtained by a random coil. A strong correlation has been observed between RPF scores of NMR structure models generated by automated NMR data analysis methods and GDT-TS/RMSD values relative to the corresponding crystal structures or manually-refined NMR structures2527.

The RPF method was adapted here as one of the assessment scores for template-based assessment in CASP10. Instead of measuring the agreement between structure models and NMR data, the modified RPF score used for CASP10 measures the agreement between the coordinates of a prediction model and an experimental (Xray or NMR) structure of the AU.

RPF scores were computed using the following procedure: From the target protein structure, a network Gtarget is built, as illustrated in Figure 1. Vertices (V) represent all N or C atoms from target structure and edges (Etarget) connect the vertices if their corresponding distance in the model structure is ≤ Dmax. A similar distance network Gmodel is built from the prediction model. The agreement between the two structures is reduced to compare the differences between the two graphs Gmodel (derived from the prediction model) and Gtarget (derived from the target structure).

An external file that holds a picture, illustration, etc.
Object name is nihms548634f1.jpg
Comparison of models using distance networks

Gtargetis generated from the target structure and Gmodel is built from a prediction model. Edges that are present in both Gtargetand Gmodel are true positives (TP). Edges present in Gtarget , but not in Gmodel are false negatives (FN). Edges that are not present in both Gtarget and Gmodel are true negatives (TN). Edges present in Gmodel , but not in Gtarget are false positives (FP). From these, the Recall R, Precision F, and F-measure F are computed as described in the Methods section.

The RPF score for comparing predicted and experimental models is described for the first time in this paper. TP, FN, FP and TN are defined in Figure 1. RPF metrics are then calculated as: Recall(model, AU) = TP/(TP+FN), which measures the percentage of close distance atom pairs from the experimental AU structure that are also close in the prediction model; Precision(model, AU) = TP/(TP+FP) which measures the percentage of close atom pairs from the prediction model that are also close in the experimental AU structure; and F-measure(model, AU) =[(1+b)*Precision(model, AU)*Recall(model, AU)]/[(b*Precision(model, AU)+Recall(model, AU)]. We use b=2 to weight the Recall higher than Precision. Operationally, we also calculate a Discriminating Power (DP score) as DP=[F(model, AU)-F(random, AU)]/[1-F(random, AU)], where F(random, AU) is the F-measure score calculated by comparing the distance networks of a random coil with the distance network of the AU structure. The distance network of a random coil is computed using atom distances expected for a freely rotating polypeptide chain model, as described by Flory and co-workers28. The Recall, Precision, and F-measure scores, as well as the normalized RPF score, the DP score, range from 0 to 1.0. The RPF score reported throughout this study is the DP score computed from the RPF algorithm.

Supplementary Figure S1 shows that at short distance cutoffs (e.g. the 5 Å cutoff used for the NMR RPF assessment score), RPF is dominated by local side-chain packing interactions information. As the distance cutoff increase, the correctness of fold starts to contribute to the RPF score. We choose distance cutoff of 9.0 Å, which seems to be a good balance of both global fold (main chain conformation) and also local side-chain and core side-chain packing information. For difficult targets with poor overall fold accuracy, the RPF score will be dominated by the main chain conformation, not the local side-chain packing. In comparing models with accurate folds, the RPF score will assign higher scores for models with better local side-chain conformations and core side-chain packing. Scores computed using a distance cutoff of 9 Å are referred to as RPF-9 scores.

If there are atoms missing from the predicted model but present in the experimental structure, interatomic distances for the missing pairs are set to 10,000 Å in the Gmodel, so as to count them as recall violations; i.e. for interatomic distances ≤ Dmax in the experimental structure, missing atoms in the predicted model were penalized in the RPF score by treating them as having distance > Dmax to all other atoms. In this way, the predicted model was penalized for the number of close distances ≤ Dmax in the experimental structure that it fails to predict, including those it fails to predict because of missing atoms. This assessment metric thus encourages groups to submit complete atomic coordinates.

The RPF measure used in CASP10 was adapted from the NMR RPF measure developed for comparing protein models against protein NMR data, which do not use oxygen atoms for the practical reason that O atoms are not observed in the NMR data. In the course of developing of RPF for CASP10, we did include all oxygen atoms for comparison, and we found that it enhanced the weight for backbone atom positions, and decreased its sensitivity to side chain core packing. In our opinion, it was therefore preferable to exclude the oxygen atoms in the CASP10 RPF-9 scores.

Assessment of Predicted Protein Models

Several metrics of protein structure accuracy were initially assessed. In keeping with the goal of encouraging predictors to submit high accuracy models, the standard GDT-TS assessment score of the LGA program20, with distance cutoffs of 1, 2, 4, and 8 Å, was replaced with the “high resolution” GDT-HA score15,29, with distance cutoffs of 0.5, 1.0, 2.0, and 4.0 Å. Besides GDT-HA, several other metrics were considered, including (i) GDC-all13, (ii) LDDT1, (iii) RPF, and (iv) Sphere Grinder (SphGr)19.

The GDT-HA and GDC-all scores are global measures of the agreement between a predicted model and the experimental structure, with GDT-HA reflecting the accuracy in placing Cα positions, and GDC-all including information about the positions of side-chain carbon atoms. On the other hand, the LDDT, RPF, and SphGr metrics are more sensitive to local structure accuracy and core packing. These intuitive perspectives were confirmed by examining many examples of superimposed models (and model fragments) with experimental structures, and comparing the corresponding scores.

Knowledge-based metrics such as ProsaII23 and MolProbity21,22 are valuable for assessing the physical reasonableness of molecular models in the absence of a “gold standard” by which to assess structural accuracy. However, it is well known that incorrect structures can have good ProsaII or MolProbity scores; e.g. a perfect alpha helix prediction can have excellent MolProbity scores even if the true structure is a beta strand. In the CASP TBM assessment, high-accuracy experimental X-ray crystal or NMR structures are available. Geometric comparisons between predicted models and these experimental structures provide a more rigorous basis for assessing the accuracy of a predicted model than knowledge-based metrics, particularly when using structural similarity metrics like GDC-all, LDDT, and RPF which include both backbone and side-chain atom positions in comparisons between the predicted model and the experimental structure. Hence, it was decided to carry out numerical assessment for ranking predicted models and predictors using only metrics that compare atomic coordinate of models with the corresponding experimental models; i.e. measures of the structural accuracy. Notwithstanding this decision, knowledge-based MolProbity scores were also computed for each predicted model and are included in the summary of assessment metrics.

Based on the considerations outlined in the preceding two paragraphs and our preliminary analyses, the numerical automated assessment of models was done using an equal weighting of 4 scores: two global alignment scores GDT-HA and GDC-all, and two superimposition-independent locally-oriented scores, RPF and LDDT, that are highly sensitive to side-chain atom positions and to the accuracy of local structure and side-chain packing. The LDDT metric was used with cutoff of 15 Å (LDDT-15), and the RPF metric was used with cutoff of 9 Å (RPF-9), as explained in the Supplementary Materials section. Our initial analyses also include SphereGrinder scores, but as omitting these scores did not have a significant impact on the rankings, our final analysis included only the four scores described above. Overall, the resulting CASP10 TBM composite assessment score is therefore quite similar to that used in CASP91, which included an equal weighting of GDT-HA, GDC-all, and LDDT-5 (5 Å cut off) scores. However, the addition of the RPF-9 score to the assessment provided more weight in the total score on local side-chain conformation and core side-chain packing relative to overall backbone structure.

Selection of Models for Assessment

As in CASP8 and CASP9, for each AU predicted by each group, only the model designated as “Model #1” (or the one with the lowest index if no “Model #1” was available) was used for automated numerical assessment. Models containing < 20 residues were excluded from assessment. For predictions that included multiple fragments, the fragment with the longest overlap with the target AU was used.

Defining Assessment Units

Assessment units were defined based on careful manual analysis of the experimental structures and potential templates, as outlined in the accompanying paper by Taylor et al17. Experimental NMR structures, as well as some X-ray crystal structures, were trimmed back to include only the consistently well-defined regions of the structure using the expanded FindCore algorithm, as described by Snyder et al18 (accompanying paper).

Numerical Automated Structure Quality Assessment Scores

All structure quality assessment scores were computed by the CASP Prediction Center19. GDT and GDC scores were computed for all predictions using the latest version of LGA19,20. The Prediction Center also provided large-scale calculations of LDDT1, RPF, Sphere Grinder19, MolProbity13,21,22, and Prosa23 scores. LDDT scores, which compare the interatomic distance matrices between a predicted model and the experimental structure1,24, were computed with a distance cutoff of 15 Å, which is larger than the 5 Å cutoff used in CASP91. They are referred to throughout this study as LDDT-15 scores. The "--rm=zeroocc" parameter was not used when computing the LDDT score. This means that atoms with zero occupancy were actually included in the LDDT score calculation. As there are not many predictions with zero occupancy atoms, this omission had minimal effect. Sphere Grinder used a 6 Å radius for the sphere, and a 2 Å rmsd cutoff. While these other metrics are described in other publications, the RPF metric for comparing predicted models against experimental structures is described in the next section.

RPF and DP Scores

The RPF method was originally developed as a quality assessment tool for protein NMR structures. The NMR RPF method calculates the Recall, Precision, F-measure and Discriminating Power (DP) scores, by measuring the agreement between all proton-proton distances ≤ 5 Å, and the NMR NOESY and chemical shift data25. Specifically, the DP score measures how well a structural model fits with the NMR data (i.e., NOESY peak list and chemical shift data), normalized by the F-measure score that would be obtained by a random coil. A strong correlation has been observed between RPF scores of NMR structure models generated by automated NMR data analysis methods and GDT-TS/RMSD values relative to the corresponding crystal structures or manually-refined NMR structures2527.

The RPF method was adapted here as one of the assessment scores for template-based assessment in CASP10. Instead of measuring the agreement between structure models and NMR data, the modified RPF score used for CASP10 measures the agreement between the coordinates of a prediction model and an experimental (Xray or NMR) structure of the AU.

RPF scores were computed using the following procedure: From the target protein structure, a network Gtarget is built, as illustrated in Figure 1. Vertices (V) represent all N or C atoms from target structure and edges (Etarget) connect the vertices if their corresponding distance in the model structure is ≤ Dmax. A similar distance network Gmodel is built from the prediction model. The agreement between the two structures is reduced to compare the differences between the two graphs Gmodel (derived from the prediction model) and Gtarget (derived from the target structure).

An external file that holds a picture, illustration, etc.
Object name is nihms548634f1.jpg
Comparison of models using distance networks

Gtargetis generated from the target structure and Gmodel is built from a prediction model. Edges that are present in both Gtargetand Gmodel are true positives (TP). Edges present in Gtarget , but not in Gmodel are false negatives (FN). Edges that are not present in both Gtarget and Gmodel are true negatives (TN). Edges present in Gmodel , but not in Gtarget are false positives (FP). From these, the Recall R, Precision F, and F-measure F are computed as described in the Methods section.

The RPF score for comparing predicted and experimental models is described for the first time in this paper. TP, FN, FP and TN are defined in Figure 1. RPF metrics are then calculated as: Recall(model, AU) = TP/(TP+FN), which measures the percentage of close distance atom pairs from the experimental AU structure that are also close in the prediction model; Precision(model, AU) = TP/(TP+FP) which measures the percentage of close atom pairs from the prediction model that are also close in the experimental AU structure; and F-measure(model, AU) =[(1+b)*Precision(model, AU)*Recall(model, AU)]/[(b*Precision(model, AU)+Recall(model, AU)]. We use b=2 to weight the Recall higher than Precision. Operationally, we also calculate a Discriminating Power (DP score) as DP=[F(model, AU)-F(random, AU)]/[1-F(random, AU)], where F(random, AU) is the F-measure score calculated by comparing the distance networks of a random coil with the distance network of the AU structure. The distance network of a random coil is computed using atom distances expected for a freely rotating polypeptide chain model, as described by Flory and co-workers28. The Recall, Precision, and F-measure scores, as well as the normalized RPF score, the DP score, range from 0 to 1.0. The RPF score reported throughout this study is the DP score computed from the RPF algorithm.

Supplementary Figure S1 shows that at short distance cutoffs (e.g. the 5 Å cutoff used for the NMR RPF assessment score), RPF is dominated by local side-chain packing interactions information. As the distance cutoff increase, the correctness of fold starts to contribute to the RPF score. We choose distance cutoff of 9.0 Å, which seems to be a good balance of both global fold (main chain conformation) and also local side-chain and core side-chain packing information. For difficult targets with poor overall fold accuracy, the RPF score will be dominated by the main chain conformation, not the local side-chain packing. In comparing models with accurate folds, the RPF score will assign higher scores for models with better local side-chain conformations and core side-chain packing. Scores computed using a distance cutoff of 9 Å are referred to as RPF-9 scores.

If there are atoms missing from the predicted model but present in the experimental structure, interatomic distances for the missing pairs are set to 10,000 Å in the Gmodel, so as to count them as recall violations; i.e. for interatomic distances ≤ Dmax in the experimental structure, missing atoms in the predicted model were penalized in the RPF score by treating them as having distance > Dmax to all other atoms. In this way, the predicted model was penalized for the number of close distances ≤ Dmax in the experimental structure that it fails to predict, including those it fails to predict because of missing atoms. This assessment metric thus encourages groups to submit complete atomic coordinates.

The RPF measure used in CASP10 was adapted from the NMR RPF measure developed for comparing protein models against protein NMR data, which do not use oxygen atoms for the practical reason that O atoms are not observed in the NMR data. In the course of developing of RPF for CASP10, we did include all oxygen atoms for comparison, and we found that it enhanced the weight for backbone atom positions, and decreased its sensitivity to side chain core packing. In our opinion, it was therefore preferable to exclude the oxygen atoms in the CASP10 RPF-9 scores.

Assessment of Predicted Protein Models

Several metrics of protein structure accuracy were initially assessed. In keeping with the goal of encouraging predictors to submit high accuracy models, the standard GDT-TS assessment score of the LGA program20, with distance cutoffs of 1, 2, 4, and 8 Å, was replaced with the “high resolution” GDT-HA score15,29, with distance cutoffs of 0.5, 1.0, 2.0, and 4.0 Å. Besides GDT-HA, several other metrics were considered, including (i) GDC-all13, (ii) LDDT1, (iii) RPF, and (iv) Sphere Grinder (SphGr)19.

The GDT-HA and GDC-all scores are global measures of the agreement between a predicted model and the experimental structure, with GDT-HA reflecting the accuracy in placing Cα positions, and GDC-all including information about the positions of side-chain carbon atoms. On the other hand, the LDDT, RPF, and SphGr metrics are more sensitive to local structure accuracy and core packing. These intuitive perspectives were confirmed by examining many examples of superimposed models (and model fragments) with experimental structures, and comparing the corresponding scores.

Knowledge-based metrics such as ProsaII23 and MolProbity21,22 are valuable for assessing the physical reasonableness of molecular models in the absence of a “gold standard” by which to assess structural accuracy. However, it is well known that incorrect structures can have good ProsaII or MolProbity scores; e.g. a perfect alpha helix prediction can have excellent MolProbity scores even if the true structure is a beta strand. In the CASP TBM assessment, high-accuracy experimental X-ray crystal or NMR structures are available. Geometric comparisons between predicted models and these experimental structures provide a more rigorous basis for assessing the accuracy of a predicted model than knowledge-based metrics, particularly when using structural similarity metrics like GDC-all, LDDT, and RPF which include both backbone and side-chain atom positions in comparisons between the predicted model and the experimental structure. Hence, it was decided to carry out numerical assessment for ranking predicted models and predictors using only metrics that compare atomic coordinate of models with the corresponding experimental models; i.e. measures of the structural accuracy. Notwithstanding this decision, knowledge-based MolProbity scores were also computed for each predicted model and are included in the summary of assessment metrics.

Based on the considerations outlined in the preceding two paragraphs and our preliminary analyses, the numerical automated assessment of models was done using an equal weighting of 4 scores: two global alignment scores GDT-HA and GDC-all, and two superimposition-independent locally-oriented scores, RPF and LDDT, that are highly sensitive to side-chain atom positions and to the accuracy of local structure and side-chain packing. The LDDT metric was used with cutoff of 15 Å (LDDT-15), and the RPF metric was used with cutoff of 9 Å (RPF-9), as explained in the Supplementary Materials section. Our initial analyses also include SphereGrinder scores, but as omitting these scores did not have a significant impact on the rankings, our final analysis included only the four scores described above. Overall, the resulting CASP10 TBM composite assessment score is therefore quite similar to that used in CASP91, which included an equal weighting of GDT-HA, GDC-all, and LDDT-5 (5 Å cut off) scores. However, the addition of the RPF-9 score to the assessment provided more weight in the total score on local side-chain conformation and core side-chain packing relative to overall backbone structure.

Selection of Models for Assessment

As in CASP8 and CASP9, for each AU predicted by each group, only the model designated as “Model #1” (or the one with the lowest index if no “Model #1” was available) was used for automated numerical assessment. Models containing < 20 residues were excluded from assessment. For predictions that included multiple fragments, the fragment with the longest overlap with the target AU was used.

Results and Discussion

As CASP has evolved there has been increasing emphasis on the accuracy of predicting realistic protein structure models1,13,29,30. The goal is to encourage CASP predictions on the TBM class of targets towards the same standards used for experimental protein structure determinations, including accurate positions for all heavy (C, N, O, and S) atoms. In CASP10 this philosophy was continued, and predictors were instructed that evaluation would include not only an assessment of the accuracy of backbone Cα atom positions, but also the accuracy of side-chain heavy atom (i.e. non-proton) positions.

Summary of Targets, Predictions and Assessment Units

During CASP 10, 114 protein sequences were released by the Prediction Center. 17 of these were later cancelled for various reasons (as outlined in Taylor et al17), leaving 97 proteins and protein domains which ranged in size from 33 residues up to 770 esidues. Following the process used in CASP7 through CASP9 3133, targets were first split into assessment units (AUs) and assigned to either the “free modeling” (FM) or “template based modeling” (TBM) assessment groups, as outlined in the accompanying paper by Taylor et al17. The TBM AUs ranged in size from 24 residues (T0709-D1) to 498 residues (T0645-D1 and T0664-D1). In total, 112 AUs associated with 91 experimental protein structures were assigned to the TBM group.

In CASP10, 41,740 predictions were submitted for the 112 assessed TBM AUs, by a total of 150 predictor groups. Of these groups, 69 were registered as prediction servers and 81 were expert “human-curated” predictor groups. The CASP process allows each submitter to provide up to 5 alternative models. However, the primary TBM assessment considered only the single model designated as Model #1; this is supposedly the best model. Hence while more than 40,000 predicted models were submitted, only 10,287 predicted models for the 112 AUs were considered in the numerical rankings. As outlined below, studies were also carried out of the impact of model selection on rankings. These hypothetical rankings required consideration of accuracy scores for all 41,740 predicted models.

As in CASP8 and CASP9, protein sequences for which templates could be easily identified by sequence-based methods were classified as “server-only targets” at the time the target was released by the CASP Prediction Center. The goal of this designation is to allow human expert groups to focus their attention on the remaining more challenging TBM and FM targets35. If a target was released as "server-only", then all AUs from this target were defined as soAUs. If a target was released as "all-group" (or human/server, hs), then all AUs from this target were hsAUs. Accordingly, 55 AUs were designated as server-only TBM AUs (soAUs). Server groups (S groups) were assessed on prediction of all 112 TBM AUs (i.e. the human or server hsAUs), while human expert groups (H groups) were assessed only on predictions for the subset of 57 TBM AUs (hAUs), excluding the 55 soAUs. Following the presentation of results for CASP8 and CASP9, the results of assessing the 112 hsAUs by the S groups, and the subset of 57 hAUs by the H and S groups, are presented separately. However, the same metrics and methods, outlined in the following section, were used to assess all predictions.

Automated Numerical Assessment

For every submitted model, the CASP10 Prediction Center19 computed raw scores for each metric and AU. These scores, for every model submitted, can be found on the CASP10 Prediction Center web site (http://www.predictioncenter.org/casp10/results.cgi). Where submitted models spanned more than one AU, numerical scores were computed relative to each AU.

The automated numerical assessment of predictions and ranking of predictor groups was carried out following the same general strategy used in previous recent CASP comparative modeling or TBM assessments1,13,15,30,36. A flow chart of the process is presented in Figure 2.

An external file that holds a picture, illustration, etc.
Object name is nihms548634f2.jpg

Flowchart of the procedure used for CASP10 TBM numerical assessment and ranking of predictor groups.

For each submitted model and corresponding AU, raw GDT-HA, GDC-all, LDDT-15, and RPF-9 scores were compiled. Next, for each AU the mean and standard deviations of these scores were computed. The mean and standard deviations for each of these distributions (for each of 112 hsAU’s and each of the four metrics) were then used to assign a Z score for each metric to each prediction model.

Ranking Based on Z Scores

As in CASP8 and CASP9, these initial Z scores were then used to eliminate the most inaccurate models for the Z score analysis. The motivation for doing this is to encourage predictors to explore new methods, and to minimize penalties that incur due to bad models that may result from the exploration of new methods37. In recent TBM CASP assessments1,15, the resulting models were then used to recalculate Z scores. As an additional motivation to minimize penalties due to poor models, models with Z scores < 0 were assigned Z = 0, and the resulting Z scores were used to compute a composite Z score for the predictor group, accounting for the performance on all models with all metrics. This composite score was then used to rank the predictor group.

As in CASP8 and CASP9, models with Z scores < −2 in first cycle of Z score analysis were excluded from the second-cycle Z score analysis (Figure 2). However, for CASP10, final Z scores after the second cycle were computed somewhat differently. The intention of assigning Z = 0 to Z scores < 0 was to encourage CASP participants to submit predictions for difficult targets. Binning the targets into groups based on their difficulty, as measured by the maximum GDT-TS value obtained by any group for that target, we observed a high frequency of negative Z scores for easier targets. For these CASP10 targets, assigning Z = 0 to Z scores < 0 had the unintended consequence of improving scores for groups than made relatively poor predictions on easy targets; i.e. AU targets for which other groups submitted accurate predictions. Interestingly, when some top scoring groups did poorly, they tended to do poorly on these easier targets; the process of assigning Z = 0 to Z scores < 0 tended to lessen the impact of these poorer models which in fact should contribute negatively to the ranking. For this CASP10 assessment, this effect was ameliorated by instead setting models with Z scores < −2 after the second cycle to Z = −2

Our analysis of the Z score distributions also revealed another caveat. In the CASP8 analysis, it was pointed out that for some AUs for which no groups provided good models, the Z score can lead to an overestimate of performance15. For the difficult AUs with best GDT-TS scores ≤ 50%, for which most predictions were quite poor, a less-poor prediction often resulted in a significantly positive Z score; i.e. the prediction was significantly better than most, but was still a very poor structure; e.g. a significantly better-than-average GDT-HA score for a poor model might result in a significantly better Z score. This effect could be addressed by including in each distribution for each metric an “ideal score”; i.e. the best score (UB) obtained for the corresponding metric for any model, on any AU, by any predictor. Figure 3 demonstrates how including a global UB raw score in the distribution used to compute Z scores for each metric suppresses high Z scores otherwise obtained for poorly-modeled difficult AUs. In the examples shown in Figures 3A and B, where a poor model prediction was substantially better than even less accurate predictions, inclusion of the UB score in the distributions suppressed the high estimate of performance by the Z score measure. Data summarized in Figure 3C demonstrate how the inclusion of UB scores reduces the high Z scores obtained for predictions of AUs for which the best GDT-TS scores (i.e. Max GDT_TS) were ≤ 50%, with less effect on Z scores of AUs for which the best GDT-TS scores were > 50%.

An external file that holds a picture, illustration, etc.
Object name is nihms548634f3.jpg
Impact on Z scores of including the global maximum raw score UB

In the example AUs shown in panels A and B, prediction models with low GDT-HA scores have high GDT-HA Z scores, even though all predictions for the corresponding AU are poor. Inclusion of the best GDT-HA score obtained for any AU (UB) in the raw score distribution, resulting in a modified Z score, Z*, suppresses the high Z scores for these poor models. A) Center – the experimental structure of AU T0739_D3. Left – model 148_1(GDT_HA=11.2 / Z=1.99 / Z*=0.513). Right – model 475_1 (GDT_HA=9.07 / Z=1.18 / Z*=0.268). B) Center - the experimental structure of T0739_D4. Left – model 112_1(GDT_HA=13.8 / Z=4.56 / Z*=0.811). Right – model 488_1 (GDT_HA=10.7 / Z=2.71 / Z*=0.445). C) Top - the Z score distributions calculated without using the UB, binned into classes of AUs based on the maximum GDT-TS (max GDT_TS score) obtained by any predictor group. For example, the left-most bin (labeled 30) presents box plots of Z scores for the AU class with max GDT_TS < 30%. Bottom – the modified Z* score distribution which includes the UB raw score the Z score analysis of each AU. Using a UB raw score in the distribution of scores for each AU significantly lowers the high Z scores otherwise obtained for poor predictions of the more difficult AUs.

With the adjustments outlined above, including a UB “best raw score” in each distribution, recomputing Z scores, and then setting Z = −2 for Z < −2, we computed Z scores for each “model #1” of each predictor group, for each of the four metrics (GDT-HA, GDC-all, LDDT-15, and RPF-9). The resulting Z scores for each metric and predictor group were then summed. Missing models were assigned a Z score of 0; generously minimizing the penalty for not submitting any prediction for an AU. These results are summarized for the 112 hsAUs in Supplementary Table S-I, and for the 57 hAUs in Supplementary Table S-II. The top 25 performing groups are listed in Table I for the server predictors and in Table II for the human curated predictors.

Table I

Sum and Average Z-scores for Top 25 Performing Server Predictor Groups – 112 hsAUs

GroupNameN_modelGDT-HAGDC-allRPFLDDTSumAvg-aAvg-sMolPro
330sBAKER-ROSETTASERVER11252.0362.4977.6075.1866.830.600.60219.08
035sZhang-Server11254.2148.4178.5266.5961.930.550.5515.62
108sPMS11237.2149.2372.7474.0858.320.520.522.03
114sQUARK11150.6841.9170.4060.6055.900.500.509.40
370sHHpred-thread11144.3948.8950.6058.0550.480.450.46−144.25
122sRaptorX-ZY11243.4442.1453.9647.2246.690.420.42−70.08
430sHHpredA11244.2046.7045.3049.2146.350.410.41−137.90
223sHHpredAQ11240.8243.8744.7348.8844.580.400.40−139.75
486sRaptorX11242.9143.5042.6236.2341.320.370.37−36.43
424sMULTICOM-NOVEL11231.3034.8845.1147.3439.660.350.3527.56
125sMULTICOM-REFINE11227.7033.0339.7442.5235.750.320.3230.97
081sMULTICOM-CLUSTER11224.3930.7239.4941.4134.000.300.3030.46
335sTASSER-VMT11229.1324.9348.7428.7532.890.290.29−88.38
103sPconsM11224.2125.5943.6137.4432.710.290.2927.14
488schunk-TASSER11224.8529.2838.7336.4232.320.290.29−28.19
292sPcons-net11213.9320.2532.8628.6623.930.210.2165.23
286sMufold-MD11210.7216.3532.4730.6322.540.200.2093.95
222sMULTICOM-CONSTRUCT11214.4019.4226.0729.1122.250.200.2016.48
333sMUFOLD-Server11216.0515.9627.7023.7920.880.190.19−13.69
261sSeok-server11213.3123.5012.7125.4018.730.170.1775.22
411sFALCON-TOPO1126.677.6717.5411.7210.900.100.10−30.77
456sFALCON-TOPO-X1123.355.1411.725.906.530.060.06−36.42
124sPconsD1110.89−0.4214.288.155.730.050.050.36
348sPhyre2_A1127.885.332.206.295.430.050.05−82.57
413sZHOU-SPARKS-X112−3.20−5.7611.418.482.730.020.02−68.11

The columns labeled GDT-HA, GDC-all, RPF, and LDDT are the sum of Z scores across all models submitted by each predictor group. The Sum column is the average of the sum of Z scores for the four metrics assessed. The Avg-s and Avg-a scores are the Sum scored divided by the number of AUs for which a model was submitted (s) by each predictor group, and the total number of AUs used for assessment (a), respectively. These scores are identical for predictor groups who submitted models for all 112 hsAUs.

Table II

Sum and Average Z-scores for Top 25 Performing Predictor Groups – 57 hAUs.

GroupNameN_modelGDT-HAGDC-allRPFLDDTSumAvg-aAvg-sMolPro
237zhang5737.6433.5050.2346.1541.880.740.741.63
27LEEcon5735.6036.1445.3947.4341.140.720.7212.84
035sZhang-Server5732.8531.0044.6840.1937.180.650.65−2.33
130Pcomb5729.9228.9944.9641.6836.390.640.6422.33
197Mufold5630.4629.7442.4139.2035.450.620.63−9.75
79TASSER5734.3432.4940.5933.9235.340.620.62−32.59
267Pcons5628.7628.7842.4840.7335.190.620.6315.26
489MULTICOM5729.8529.5340.1739.0234.640.610.619.01
344Jones-UCL5626.7828.7243.0837.8234.100.600.61−50.10
114sQUARK5629.6825.9838.5035.8732.510.570.58−5.65
301LEE5727.5329.3834.9837.5232.350.570.576.37
477BAKER5731.4629.2435.3433.0432.270.570.57104.97
475CNIO5727.8623.5239.7837.6932.210.570.57−2.34
350Kloczkowski_Lab5726.7423.8240.0136.9431.880.560.5618.79
490Zhang_Refinement5726.0324.8133.2532.6129.180.510.5116.67
294chuo-repack5723.1920.5536.7934.1928.680.500.50−6.90
458St ernberg5725.9423.3032.2531.6628.290.500.50−10.66
365chuo-fams5722.5021.0835.9231.9227.860.490.494.58
428PconsQ5622.0820.0933.9233.4327.380.480.497.49
434chuo-fams-consensus5720.1915.1632.6831.4124.860.440.44−7.49
481Chicken_George5721.7018.5031.0228.1724.850.440.445.58
122sRaptorX-ZY5725.8824.3424.6821.9824.220.430.43−44.38
45Zhang_Ab_Initio5721.0720.3728.2826.9624.170.420.424.98
285McGuffin5518.2816.2630.5227.9223.250.410.42−1.60
405Mufold25418.9818.3628.9225.9323.050.400.43−17.03

The columns labeled GDT-HA, GDC-all, RPF, and LDDT are the sum of Z scores across all models submitted by each predictor group. The Sum column is the average of the sum of Z scores for the four metrics assessed. The Avg-s and Avg-a scores are the Sum scored divided by the number of AUs for which a model was submitted (s) by each predictor group, and the total number of AUs used for assessment (a), respectively. These scores are identical for predictor groups who submitted models for all 57 hAUs.

The sum of model Z scores for each metric by each group (Tables I and andII)II) provides a useful assessment of performance with respect to that particular metric. However, for ranking the performance of the predictor groups it is useful to have a single composite performance score. For CASP10, the composite performance score (Sums in Table I and Table II) was computed by averaging, for each predictor group, the summed Z scores for the four metrics.

An average performance score per AU was then calculated by dividing this composite score by the number of AUs. This average performance score for each group was computed in two ways: (i) dividing the composite performance score by the total number of available AUs (Avg-a) (i.e. 112 or 57), or (ii) dividing by the total number of AUs submitted by the group (Avg-s). The former method penalizes groups that did not submit predictions for all AUs. For the top performing groups, all of which provided predictions for 111–112 hsAUs (Table I) [or for 56–57 of the hAUs (Table II)], these average values were approximately the same. The rankings shown in Tables I and andIIII are based on the sum average Avg-a, which slightly penalizes the groups that did not submit predictions for all of the available AUs.

The sum MolProbity21,22 Z scores for each predictor group is also tabulated for the top 25 performing groups in Tables I and andIIII (and in Supplementary Tables I and II for all groups). Although these MolProbity scores were not used in ranking predictor groups, they provide a knowledge-based assessment of structure quality. In particular, the Baker group consistently provided models with good MolProbity scores.

Head-to-head Paired Student’s t-test

The Z score ranking was not used to provide a final ranking of predictor groups. As in CASP91, the Z score ranking (based on the sum average Avg-a scores) was only used to identify the top 25 groups, which were then selected for a more detailed assessment accounting for the statistical significance of ranking one group above another. The predictions of these 25 groups were then compared in a direct head-to-head statistical analysis on common targets. In this analysis, each score distribution (GDT-HA, GDC-all, LDDT-15, RPF-9) was considered separately. Using each of these metrics independently, the raw score distributions for each of the AUs from each predictor group were compared with the corresponding score distributions from other groups in the top 25 list, using pair-wise Student’s t-test. The number of comparisons for which there was a statistically significant difference was then summed for each metric. For each score metric, each of the top 25 groups is compared with 24 other groups. Group A was assigned 1 point if its distribution of raw scores for that metric is significantly better (p < 0.05) than that of group B. Hence, the maximum score per metric is 24; and the maximum summed score for four metrics assessed is 96. For models with Z scores < −2, the raw score was set to the raw score value for that AU corresponding to Z = −2 (based on the second cycle Z score analyses summarized in Tables I and andII).II). This consideration was relatively insignificant for most of the 25 top-performing groups, for which even the worst models generally have Z > −2. When a model was missing for a group, the median score for that measure was used as the raw score. This consideration was also insignificant, as most of the 25 top-performing groups submitted prediction models for all AU targets. These summed Student’s t-test head-to-head pairwise scores (H2H Scores), plotted in Figure 4A (for the 112 hsAUs) and Figure 4B (for the 57 hAUs), provided the basis for determining the final ranking of predictor group.

An external file that holds a picture, illustration, etc.
Object name is nihms548634f4.jpg
Ranking of top 25 server-only and human / server predictor groups

(A) Head-to-head pairwise Student’s t-test analysis on raw scores between 25 top-ranking server predictor groups for 112 hsAUs. (B) Head-to-head pairwise Student’s t-test analysis on raw scores between 25 top-ranking server / human predictor groups for 57 hAUs. Black, GDT-HA; red, GDC-all; green, RPF-9; blue, LDDT-15. Top-ranking 25 groups were identified based on average Z score Avg-a (Tables I and andIIII).

It is important to recognize that the final ranking among the top 25 performing groups did not depend on the details of our methods for computing Z scores; the Z scores were used only to provide an overall ranking of all groups (presented in Supplementary Tables I and II), and to identify the top 25 performing groups. The ranking within these top 25 groups was based on the raw scores for each of the four metrics, using paired Student’s t-test in head-to-head comparisons of the predictions made by each group on the accuracy of the predictions of the same target by all of the other 24 groups. In all of the H2H analyses presented in this paper, the order of the groups shown in the H2H analysis plots for which the distribution of raw prediction scores for each of the four metrics assessed were not statistically better than any of the other top 24 groups, are random. The relative rankings of these groups are determined by the Z score statistics of Supplementary Tables S-I, S-II, and S-III.

H2H Ranking – Servers

Figure 4A summarizes the results of the head-to-head (H2H) Student’s t-test analysis for the top 25 performing server groups, assessed on 112 AUs. The five top performing groups were Zhang Server (035s), QUARK (114s), PMS (108s), HHpred-thread (370s), and RaptorX-ZY(122s). These were followed by BAKER-ROSETTASERVER (330s), HHPredA (430s), RaptorX (486s), HHPredAQ (223s), and MULTICOM-NOVEL (424s).

Generally, the H2H ranking of Figure 4A correlates with the Z score ranking of Table I: e.g. 330s > 035s > 108s > 114s > 370s > 122s > 430s > 223s > 486s > 424s. One notable outlier in this correlation is the BAKER-ROSETTASERVER (330s), which was top-ranked in the Z score analysis (Table I), but sixth-ranked in the H2H ranking (Figure 4A). Figure 5 compares GDT-HA raw score and Z score distributions for predictions made by Zhang Server (035s) and BAKER-ROSETTASERVER (330s). This analysis demonstrates that, based on raw GDT-HA scores (Figure 5A), 035s had better performance on the generally more challenging hAUs (i.e. targets with low sequence identity with templates) than 330s. Performance by 330s was marginally better on soAUs, with easily identified templates of known structure. These results suggest that 035s was able to do a superior job than 330s in identifying distant templates. Although this same better performance on more challenging hAUs is evident in the Z score analysis (Figure 4B), these Z scores are much less discriminating than the raw scores, particularly for the challenging hAUs with very low sequence identity with templates. The classification of soAUs and hAUs was made by the CASP Prediction Center based on the HHSearch probability score with potential templates, and may not always accurately reflect the actual difficulty of the targets. In addition, we observed many examples where Z scores overestimate prediction performance when the spreads of the metric for some specific targets are narrow (Supplementary Figure S6). Overall, these results demonstrate the value of using the raw accuracy metric scores, through the H2H paired Student’s t-test, in determining the final ranking.

An external file that holds a picture, illustration, etc.
Object name is nihms548634f5.jpg
Raw metric scores provide more sensitivity to poor performance on more challenging hAUs

Comparison of performance by top-ranking groups shows how (A) GDT-HA raw scores are more sensitive to poor performance on more challenging hAUs than (B) GDT-HA Z scores. Close circles, hAU targets. Open circles, soAU targets. The three dashed lines of Panel A are at y = × and y = × ± 20 units, and the three dashed lines of Panel B are at y = × and y = × ± 1.5 units. Panel B is expanded to exclude regions of the plot with no data points.

We also explored the sensitivity of the H2H ranking against some extremely poor predictions. One approach for suppressing the high impact of those extreme outliers is to replace the Student’s t-test with the non-parametric Wilcoxon signed rank test. This statistic compares the rank, rather than value, of the raw score differences between two distributions, and is less sensitive to raw score outliers. As illustrated in Supplementary Figure S2, suppression of outliers using the pairwise Wilcoxon rank sum analysis alters the ranking among the top 25 performing groups; e.g. in this analysis BAKER-ROSETTASERVER (330s) becomes the top ranked server.

H2H Ranking – Human and/or Servers

Figure 4B summarizes the results of the head-to-head (H2H) paired Student’s t-test analysis for the top 25 performing human and/or server groups, assessed on the 57 hAUs. The score for the top-performing group, Zhang (237), was significantly higher than any other group. The next best performing groups were Leecon (027) and Zhang-Server (035s), followed by TASSER (079), Pcomb (130), Pcons (267), QUARK (114s), MULTICOM (489), CNIO (475), and Mufold (197).

For the Human and/or Server analysis, the H2H ranking was generally well-correlated with the Z score ranking (Table II); 237 > 027 > 035s > 130 > 197 > 79 > 267 > 489 > 344 > 114s. The differences between H2H and Z score ranking again could be traced to the higher sensitivity of raw score comparisons over Z score comparison, especially for hAUs with low sequence identity with templates.

H2H Ranking – Most Difficult TBMs

We also ranked the human and/or server predictors using only the 15 “most difficult” AUs. Figure 6 is a bar plot of the maximum GDT-TS (Max GDT-TS) score obtained for each of the 112 hsAU’s by any of the predictors. AUs with Max GDT-TS ≤ 50% are referred to here as TBM_hard AUs. Based on this analysis, 15 CASP10 AUs are defined as the “TBM_hard AUs”: T0649_D1, T0653_D1, T0668_D1, T0671_D2, T0676_D1, T0678_D1, T0684_D1, T0690_D1, T0705_D2, T0717_D2, T0726_D1, T0732_D2, T0735_D1, T0739_D3, T0739_D4. Using only these 15 targets, we repeated the Z score analysis, shown in Table III for the top-ranked 25 groups, and in Supplementary Table S-III for all groups. We also carried out head-to-head paired Student’s t-test analysis of raw scores among the top-performing 25 groups. Figure 7 summarizes the results of the H2H, assessed on only the 15 TBM_hard AUs. For this set of targets, the top-performing groups were zhang (237) and MULTICOM (489), with nearly identical performance that was significantly better than that for the other 23 groups. These were followed by MuFold (197), TASSER (079), Pcons (267), chuo-fams (365), RaptorX-ZY (122s), Pcomb (130), Kloczkowski (350) and Leecon (027). Significantly, of the 69 server predictors participating in CASP10, only one server, RaptorX-ZY (122s), performed well with these TBM_hard AUs.

An external file that holds a picture, illustration, etc.
Object name is nihms548634f6.jpg
Classifying TBM_hard AUs

Bar plot of the maximum GDT-TS (Max GDT-TS) score obtained for each of the 112 AU’s (indicated by AU target id) by any of the predictors. The 15 AU targets with Max GDT-TS ≤ 50% were identified as TBM_hard AUs.

An external file that holds a picture, illustration, etc.
Object name is nihms548634f7.jpg
Ranking of top 25 human / server predictor groups on TBM_hard targets

Head-to-head (H2H) pairwise Student’s t-test analysis on raw scores between 25 top-ranking server/human predictor groups for 15 TBM_hard AUs. Black, GDT-HA; red, GDC-all; green, RPF-9; blue, LDDT-15. Top-ranking 25 groups were identified based the Avg-a Z score.

Table III

Sum and Average Z-scores for Top 25 Performing Predictor Groups – 15 TBM_Hard AUs.

GroupNameN_modelGDT-HAGDC-allRPFLDDTSumAvg-aAvg-sMolPro
489MULTICOM157.246.5814.7611.9910.140.680.681.48
237zhang156.844.9512.4610.988.810.590.59−0.84
197Mufold146.023.9212.0910.408.110.540.58−6.17
267Pcons155.284.3911.3710.377.850.520.521.74
428PconsQ154.914.0410.909.807.410.490.49−0.36
365chuo-fams154.682.8811.7710.177.380.490.491.99
27LEEcon155.083.3810.559.447.110.470.473.17
130Pcomb154.553.0511.109.176.970.460.468.60
035sZhang-Server155.183.559.858.886.870.460.46−3.49
344Jones-UCL144.303.4710.258.546.640.440.47−13.90
294chuo-repack153.432.4410.629.406.470.430.43−3.02
350Kloczkowski_Lab153.182.3710.849.006.350.420.42−1.04
285McGuffin144.133.159.137.776.050.400.43−0.45
475CNIO153.551.799.998.625.990.400.40−1.58
79TASSER155.566.947.173.985.910.390.39−2.92
434chuo-fams-consensus152.831.559.518.675.640.380.38−0.49
114sQUARK143.832.738.457.315.580.370.40−5.13
315keasar153.882.249.026.395.380.360.36−15.88
122sRaptorX-ZY155.885.625.563.755.200.350.35−14.48
45Zhang_Ab_Initio152.963.377.686.715.180.350.351.19
490Zhang_Refinement154.002.567.526.405.120.340.344.24
26Pr oQ2clust142.783.416.796.684.920.330.35−1.75
101WeFold142.942.388.165.634.780.320.34−9.59
388ProQ2152.111.118.257.484.740.320.3220.29
330sBAKER-ROSETTASERVER152.632.577.265.784.560.300.3029.53

The columns labeled GDT-HA, GDC-all, RPF, and LDDT are the sum of Z scores across all models submitted by each predictor group. The Sum column is the average of the sum of Z scores for the four metrics assessed. The Avg-s and Avg-a scores are the Sum scored divided by the number of AUs for which a model was submitted (s) by each predictor group, and the total number of AUs used for assessment (a), respectively. These scores are identical for predictor groups who submitted models for all TBM_Hard hAUs.

It was particularly important to use multiple metrics of model accuracy in assessing performance on the TBM_hard AUs, as some groups scored well with some metrics, other groups scored well with different metrics. Only the two top-ranked groups zhang (237) and MULTICOM (489) consistently scored well with all four metrics. This is distinct from the results with 112 hsAUs or 57 hAUs, in which there was generally consistent performance on all four metrics by the top scoring groups.

Impact of Distance Matrix Methods in Assessing Low Accuracy Structures

Figure 7 shows that the superimposition-independent LDDT-15 and RPF-9 scores are helpful in distinguishing performance of predictor groups on the TBM_hard AU’s. Also relevant was the complementary value of LDDT-15 and RPF-9 scores in ranking TBM_hard AUs; some predictor groups did best with LDDT-15 scores, while other did best with RFP-9 scores.

Unlike other metrics used in CASP10, the RPF-9 score is normalized against a free rotating chain model, based on Flory polymer chain statistics25,28. For this reason, RPF-9 DP scores used in CASP10 are very discriminative against random structures. Random-like incorrect prediction models (i.e. low quality structures) will have RPF-9 scores very close to zero. Structures with incorrect secondary structures or incorrect folds can even have negative RPF-9 scores, indicating that they are even worse than random structures. Examples discussed in the Supplemental Material demonstrate that the RPF-9 scores normalized to random structure (i.e. the RPF-9 DP scores) have stronger discriminating power than the GDT_HA scores against structures with random-like incorrect folds (Supplementary Figure S3). Normalization against a set of well-defined decoy models, representing a random distribution of structures, has also recently been introduced into the LDDT score24. Additional examples comparing the sensitivity of global alignment (GDT and GDC-all) scores and superimposition-independent (LDDT and RPF) scores for assessing surface loop and interhelical packing inaccuracies are also presented in the Supplementary Material (Supplementary Figure S4).

Figure 7 also shows that TASSER (079), RaptorX-ZY (122s), and MULTICOM (489) had the best performance on TBM_hard AUs with the GDC-all score, although the first two of these groups did not do so well with other three scores. 3rd ranked Mufold (197) had a zero GDC-all count. It appears that TASSER and RaptorX-ZY may have some part of the core packing matched with some experimental structures, yielding higher GDC-all scores even though the main chains of these models may not be more accurate than other models from other groups. However, global alignment scores like GDC-all are challenged when the overall backbone structure is inaccurate, suggesting that GDT-HA and GDC-all scores may be unreliable for assessing predictions of difficult TBM-hard AUs. Despite some of the valuable features of the superimposition-independent metrics like RPF and LDDT for difficult targets like TBM-hard AUs, human judgment like that used in the CASP10 FM assessment14 is still a better approach than using automated scoring metrics.

Impact of accurate model selection

As discussed above in the H2H assessment of severs, a few poor models can make a significant impact on the relative ranking, particularly in distinguishing the top-performing group from other groups that generally did very well in CASP10. In many cases, predictor groups included in their five submitted models more accurate models than the one designated as Model 1. This is a well-recognized aspect of previous CASP experiments. Interestingly, our analysis suggests that at least in some cases, the more accurate model could have been selected using alternative fold-accuracy discriminators. Figure 8 illustrates two examples, one for server BAKER-ROSETTASERVER (330s) and one for human-curated group Sternberg (458) where an alternative model is more accurate (higher GDT-HA raw score) than the designated Model #1. In these two cases, the more accurate model also has a significantly better ProsaII23 fold score.

An external file that holds a picture, illustration, etc.
Object name is nihms548634f8.jpg
The challenge of accurate model selection

Predictor groups often submitted a much less accurate model as Model #1 compared to other of their own submitted alternative models. (A) For AU target T0742_D1, Model #1 submitted by group 035s (Zhang-Server) has much higher GDT-HA score (∼ 0.3; y-axis) than Model #1 submitted by server group 330s (BAKER-ROSETTASERVER; GDT-HA ∼ 0.05). However, alternative Model #3 of group 330s (GDT-HA ∼0.28) is a more accurate model. (B) For AU target T0726_D2, Model #1 submitted by the human-curated group 237 (zhang) group had much higher GDT-HA score (∼ 0.72) than Model #1 submitted by group 458 (Sternberg; GDT-HA ∼ 0.12)). However, alternative Model #2 of group 458 (GDT-HA ∼ 0.71) is a much more accurate model. In these particular cases, the Prosa Z score (x-axis) could have provided a good criteria in model selecting, as it would have indicated a much more accurate “best model” for groups 330s and 458.

We also did a comparison of GDT-HA scores between models picked from among submitted models by ProsaII score only, and the Model #1 provided by predictors for all 10,902 combinations of Groups and AUs for which ProsaII scores are available from the Prediction Center. If all the groups used ProsaII to pick their best model from the five submitted, 91 groups would improve their average GDT-HA score. Our point is not that ProsaII is the single best indicator of model quality. However, as is generally appreciated, many CASP10 participants would benefit by more successful model selection.

Ranking Based on Ideal Model Selection

Selection of the most-accurate model among submitted alternate models is an important area for development in order to provide more accurate TBM predictions. These trends were validated by simulating an “ideal model selection” CASP10 competition, choosing for each group and AU the single model with highest GDT-HA as Model #1, and repeating the Z score and H2H raw score rankings as outlined in the previous sections. Rankings based on H2H Student’s t-test raw score comparisons are shown in Figure 9A for server predictors and Figure 9B for human-curated predictors. While “ideal model selection” has modest impact on server rankings, it has a significant impact on human-curator predictor rankings (cf. the ranking of predictors in Figures 4 and and9).9). Server predictors for which ranking was significantly improved in this simulated ideal model selection CASP10 competition included BAKER-ROSETTASERVER (330s), PconsM (103s), MULTICOM-CLUSTER (081S), and MULTICOM-REFINE (125s). Human-curated predictors for which ranking was significantly improved by simulated ideal model selection include CNIO (475) (which using ideal model selection ranked second behind Zhang (237), Baker (477), chuo-repack (294), chuo-fams (365), Zhang-refinement (490), and keasar (315). These results demonstrate the key role of model selection in accurate TBM protein structure prediction.

An external file that holds a picture, illustration, etc.
Object name is nihms548634f9.jpg
Ideal model selection

For each predictor group, the submitted model with highest GDT-HA score was selected and the paired Student’s t-test analysis was done as in Fig. 4. (A) Head-to-head pairwise Student’s t-test analysis on raw scores between 25 top-ranking server predictor groups for 112 hsAUs. (B) Head-to-head pairwise Student’s t-test analysis on raw scores between 25 top-ranking server / human predictor groups for 57 hAUs. Black, GDT-HA; red, GDC-all; green, RPF-9; blue, LDDT-15. Top-ranking 25 groups were identified based the Avg-a Z score.

Conclusions

The CASP10 TBM assessment identified the Zhang-Server (035s), QUARK (114s), PMS (108s), Leecon (027), Zhang (237) groups as providing the most accurate models for the AU targets. In the course of our assessment, we observed many examples where raw scores were more discriminating for distinguishing structural accuracy than Z scores (also see Supplementary Figure S6), or where Z scores suggest good prediction performance while the raw scores together with visual inspecting reveal inaccurate models (e.g. Figure 3). In our experience, rankings based on raw scores, using head-to-head (H2H) comparisons of performance on common AUs, with a statistical assessment of whether one group outperforms each of the other groups, provides a more sensitive automated numerical assessment than Z scores alone. The H2H ranking developed by Mariani et al1 incorporates Student’s t-test, providing as statistically sound basis for ranking the relative performance of groups on the same set of targets.

The CASP10 TBM assessment, evolving from the previous CASP assessment protocols, put heavy emphasis on side-chain atom positions, which are assessed by three (GDC-all, LDDT-15, and RPF-9) of the four structure accuracy metrics that contribute to our composite score. The philosophy was to assess TBM models with the same kinds of expectations for physically-reasonable structures as is expected for experimental NMR or X-ray crystal structures. We also combined both global alignment scores (GDT-HA and GDC-all) which utilized superimpositions, and distance matrix based methods (LDDT-15 and RPF-9) which do not involve superimposition methods and are more sensitive to local structure accuracy and core side-chain packing features. The distance matrix based methods seem to be less sensitive to the loosely packed loop regions and also to helix tilt angles. The introduction of normalizing the RPF score against a random coil model impacts the ranking for the difficult targets. These multiple scores incorporate information on all heavy atom (C, N, O, and S) positions, and balance the biases inherent in superimposition-based structural comparisons (e.g. GDT scores) which are problematic when comparing structures lacking high overall structural similarity.

Summary of Targets, Predictions and Assessment Units

During CASP 10, 114 protein sequences were released by the Prediction Center. 17 of these were later cancelled for various reasons (as outlined in Taylor et al17), leaving 97 proteins and protein domains which ranged in size from 33 residues up to 770 esidues. Following the process used in CASP7 through CASP9 3133, targets were first split into assessment units (AUs) and assigned to either the “free modeling” (FM) or “template based modeling” (TBM) assessment groups, as outlined in the accompanying paper by Taylor et al17. The TBM AUs ranged in size from 24 residues (T0709-D1) to 498 residues (T0645-D1 and T0664-D1). In total, 112 AUs associated with 91 experimental protein structures were assigned to the TBM group.

In CASP10, 41,740 predictions were submitted for the 112 assessed TBM AUs, by a total of 150 predictor groups. Of these groups, 69 were registered as prediction servers and 81 were expert “human-curated” predictor groups. The CASP process allows each submitter to provide up to 5 alternative models. However, the primary TBM assessment considered only the single model designated as Model #1; this is supposedly the best model. Hence while more than 40,000 predicted models were submitted, only 10,287 predicted models for the 112 AUs were considered in the numerical rankings. As outlined below, studies were also carried out of the impact of model selection on rankings. These hypothetical rankings required consideration of accuracy scores for all 41,740 predicted models.

As in CASP8 and CASP9, protein sequences for which templates could be easily identified by sequence-based methods were classified as “server-only targets” at the time the target was released by the CASP Prediction Center. The goal of this designation is to allow human expert groups to focus their attention on the remaining more challenging TBM and FM targets35. If a target was released as "server-only", then all AUs from this target were defined as soAUs. If a target was released as "all-group" (or human/server, hs), then all AUs from this target were hsAUs. Accordingly, 55 AUs were designated as server-only TBM AUs (soAUs). Server groups (S groups) were assessed on prediction of all 112 TBM AUs (i.e. the human or server hsAUs), while human expert groups (H groups) were assessed only on predictions for the subset of 57 TBM AUs (hAUs), excluding the 55 soAUs. Following the presentation of results for CASP8 and CASP9, the results of assessing the 112 hsAUs by the S groups, and the subset of 57 hAUs by the H and S groups, are presented separately. However, the same metrics and methods, outlined in the following section, were used to assess all predictions.

Automated Numerical Assessment

For every submitted model, the CASP10 Prediction Center19 computed raw scores for each metric and AU. These scores, for every model submitted, can be found on the CASP10 Prediction Center web site (http://www.predictioncenter.org/casp10/results.cgi). Where submitted models spanned more than one AU, numerical scores were computed relative to each AU.

The automated numerical assessment of predictions and ranking of predictor groups was carried out following the same general strategy used in previous recent CASP comparative modeling or TBM assessments1,13,15,30,36. A flow chart of the process is presented in Figure 2.

An external file that holds a picture, illustration, etc.
Object name is nihms548634f2.jpg

Flowchart of the procedure used for CASP10 TBM numerical assessment and ranking of predictor groups.

For each submitted model and corresponding AU, raw GDT-HA, GDC-all, LDDT-15, and RPF-9 scores were compiled. Next, for each AU the mean and standard deviations of these scores were computed. The mean and standard deviations for each of these distributions (for each of 112 hsAU’s and each of the four metrics) were then used to assign a Z score for each metric to each prediction model.

Ranking Based on Z Scores

As in CASP8 and CASP9, these initial Z scores were then used to eliminate the most inaccurate models for the Z score analysis. The motivation for doing this is to encourage predictors to explore new methods, and to minimize penalties that incur due to bad models that may result from the exploration of new methods37. In recent TBM CASP assessments1,15, the resulting models were then used to recalculate Z scores. As an additional motivation to minimize penalties due to poor models, models with Z scores < 0 were assigned Z = 0, and the resulting Z scores were used to compute a composite Z score for the predictor group, accounting for the performance on all models with all metrics. This composite score was then used to rank the predictor group.

As in CASP8 and CASP9, models with Z scores < −2 in first cycle of Z score analysis were excluded from the second-cycle Z score analysis (Figure 2). However, for CASP10, final Z scores after the second cycle were computed somewhat differently. The intention of assigning Z = 0 to Z scores < 0 was to encourage CASP participants to submit predictions for difficult targets. Binning the targets into groups based on their difficulty, as measured by the maximum GDT-TS value obtained by any group for that target, we observed a high frequency of negative Z scores for easier targets. For these CASP10 targets, assigning Z = 0 to Z scores < 0 had the unintended consequence of improving scores for groups than made relatively poor predictions on easy targets; i.e. AU targets for which other groups submitted accurate predictions. Interestingly, when some top scoring groups did poorly, they tended to do poorly on these easier targets; the process of assigning Z = 0 to Z scores < 0 tended to lessen the impact of these poorer models which in fact should contribute negatively to the ranking. For this CASP10 assessment, this effect was ameliorated by instead setting models with Z scores < −2 after the second cycle to Z = −2

Our analysis of the Z score distributions also revealed another caveat. In the CASP8 analysis, it was pointed out that for some AUs for which no groups provided good models, the Z score can lead to an overestimate of performance15. For the difficult AUs with best GDT-TS scores ≤ 50%, for which most predictions were quite poor, a less-poor prediction often resulted in a significantly positive Z score; i.e. the prediction was significantly better than most, but was still a very poor structure; e.g. a significantly better-than-average GDT-HA score for a poor model might result in a significantly better Z score. This effect could be addressed by including in each distribution for each metric an “ideal score”; i.e. the best score (UB) obtained for the corresponding metric for any model, on any AU, by any predictor. Figure 3 demonstrates how including a global UB raw score in the distribution used to compute Z scores for each metric suppresses high Z scores otherwise obtained for poorly-modeled difficult AUs. In the examples shown in Figures 3A and B, where a poor model prediction was substantially better than even less accurate predictions, inclusion of the UB score in the distributions suppressed the high estimate of performance by the Z score measure. Data summarized in Figure 3C demonstrate how the inclusion of UB scores reduces the high Z scores obtained for predictions of AUs for which the best GDT-TS scores (i.e. Max GDT_TS) were ≤ 50%, with less effect on Z scores of AUs for which the best GDT-TS scores were > 50%.

An external file that holds a picture, illustration, etc.
Object name is nihms548634f3.jpg
Impact on Z scores of including the global maximum raw score UB

In the example AUs shown in panels A and B, prediction models with low GDT-HA scores have high GDT-HA Z scores, even though all predictions for the corresponding AU are poor. Inclusion of the best GDT-HA score obtained for any AU (UB) in the raw score distribution, resulting in a modified Z score, Z*, suppresses the high Z scores for these poor models. A) Center – the experimental structure of AU T0739_D3. Left – model 148_1(GDT_HA=11.2 / Z=1.99 / Z*=0.513). Right – model 475_1 (GDT_HA=9.07 / Z=1.18 / Z*=0.268). B) Center - the experimental structure of T0739_D4. Left – model 112_1(GDT_HA=13.8 / Z=4.56 / Z*=0.811). Right – model 488_1 (GDT_HA=10.7 / Z=2.71 / Z*=0.445). C) Top - the Z score distributions calculated without using the UB, binned into classes of AUs based on the maximum GDT-TS (max GDT_TS score) obtained by any predictor group. For example, the left-most bin (labeled 30) presents box plots of Z scores for the AU class with max GDT_TS < 30%. Bottom – the modified Z* score distribution which includes the UB raw score the Z score analysis of each AU. Using a UB raw score in the distribution of scores for each AU significantly lowers the high Z scores otherwise obtained for poor predictions of the more difficult AUs.

With the adjustments outlined above, including a UB “best raw score” in each distribution, recomputing Z scores, and then setting Z = −2 for Z < −2, we computed Z scores for each “model #1” of each predictor group, for each of the four metrics (GDT-HA, GDC-all, LDDT-15, and RPF-9). The resulting Z scores for each metric and predictor group were then summed. Missing models were assigned a Z score of 0; generously minimizing the penalty for not submitting any prediction for an AU. These results are summarized for the 112 hsAUs in Supplementary Table S-I, and for the 57 hAUs in Supplementary Table S-II. The top 25 performing groups are listed in Table I for the server predictors and in Table II for the human curated predictors.

Table I

Sum and Average Z-scores for Top 25 Performing Server Predictor Groups – 112 hsAUs

GroupNameN_modelGDT-HAGDC-allRPFLDDTSumAvg-aAvg-sMolPro
330sBAKER-ROSETTASERVER11252.0362.4977.6075.1866.830.600.60219.08
035sZhang-Server11254.2148.4178.5266.5961.930.550.5515.62
108sPMS11237.2149.2372.7474.0858.320.520.522.03
114sQUARK11150.6841.9170.4060.6055.900.500.509.40
370sHHpred-thread11144.3948.8950.6058.0550.480.450.46−144.25
122sRaptorX-ZY11243.4442.1453.9647.2246.690.420.42−70.08
430sHHpredA11244.2046.7045.3049.2146.350.410.41−137.90
223sHHpredAQ11240.8243.8744.7348.8844.580.400.40−139.75
486sRaptorX11242.9143.5042.6236.2341.320.370.37−36.43
424sMULTICOM-NOVEL11231.3034.8845.1147.3439.660.350.3527.56
125sMULTICOM-REFINE11227.7033.0339.7442.5235.750.320.3230.97
081sMULTICOM-CLUSTER11224.3930.7239.4941.4134.000.300.3030.46
335sTASSER-VMT11229.1324.9348.7428.7532.890.290.29−88.38
103sPconsM11224.2125.5943.6137.4432.710.290.2927.14
488schunk-TASSER11224.8529.2838.7336.4232.320.290.29−28.19
292sPcons-net11213.9320.2532.8628.6623.930.210.2165.23
286sMufold-MD11210.7216.3532.4730.6322.540.200.2093.95
222sMULTICOM-CONSTRUCT11214.4019.4226.0729.1122.250.200.2016.48
333sMUFOLD-Server11216.0515.9627.7023.7920.880.190.19−13.69
261sSeok-server11213.3123.5012.7125.4018.730.170.1775.22
411sFALCON-TOPO1126.677.6717.5411.7210.900.100.10−30.77
456sFALCON-TOPO-X1123.355.1411.725.906.530.060.06−36.42
124sPconsD1110.89−0.4214.288.155.730.050.050.36
348sPhyre2_A1127.885.332.206.295.430.050.05−82.57
413sZHOU-SPARKS-X112−3.20−5.7611.418.482.730.020.02−68.11

The columns labeled GDT-HA, GDC-all, RPF, and LDDT are the sum of Z scores across all models submitted by each predictor group. The Sum column is the average of the sum of Z scores for the four metrics assessed. The Avg-s and Avg-a scores are the Sum scored divided by the number of AUs for which a model was submitted (s) by each predictor group, and the total number of AUs used for assessment (a), respectively. These scores are identical for predictor groups who submitted models for all 112 hsAUs.

Table II

Sum and Average Z-scores for Top 25 Performing Predictor Groups – 57 hAUs.

GroupNameN_modelGDT-HAGDC-allRPFLDDTSumAvg-aAvg-sMolPro
237zhang5737.6433.5050.2346.1541.880.740.741.63
27LEEcon5735.6036.1445.3947.4341.140.720.7212.84
035sZhang-Server5732.8531.0044.6840.1937.180.650.65−2.33
130Pcomb5729.9228.9944.9641.6836.390.640.6422.33
197Mufold5630.4629.7442.4139.2035.450.620.63−9.75
79TASSER5734.3432.4940.5933.9235.340.620.62−32.59
267Pcons5628.7628.7842.4840.7335.190.620.6315.26
489MULTICOM5729.8529.5340.1739.0234.640.610.619.01
344Jones-UCL5626.7828.7243.0837.8234.100.600.61−50.10
114sQUARK5629.6825.9838.5035.8732.510.570.58−5.65
301LEE5727.5329.3834.9837.5232.350.570.576.37
477BAKER5731.4629.2435.3433.0432.270.570.57104.97
475CNIO5727.8623.5239.7837.6932.210.570.57−2.34
350Kloczkowski_Lab5726.7423.8240.0136.9431.880.560.5618.79
490Zhang_Refinement5726.0324.8133.2532.6129.180.510.5116.67
294chuo-repack5723.1920.5536.7934.1928.680.500.50−6.90
458St ernberg5725.9423.3032.2531.6628.290.500.50−10.66
365chuo-fams5722.5021.0835.9231.9227.860.490.494.58
428PconsQ5622.0820.0933.9233.4327.380.480.497.49
434chuo-fams-consensus5720.1915.1632.6831.4124.860.440.44−7.49
481Chicken_George5721.7018.5031.0228.1724.850.440.445.58
122sRaptorX-ZY5725.8824.3424.6821.9824.220.430.43−44.38
45Zhang_Ab_Initio5721.0720.3728.2826.9624.170.420.424.98
285McGuffin5518.2816.2630.5227.9223.250.410.42−1.60
405Mufold25418.9818.3628.9225.9323.050.400.43−17.03

The columns labeled GDT-HA, GDC-all, RPF, and LDDT are the sum of Z scores across all models submitted by each predictor group. The Sum column is the average of the sum of Z scores for the four metrics assessed. The Avg-s and Avg-a scores are the Sum scored divided by the number of AUs for which a model was submitted (s) by each predictor group, and the total number of AUs used for assessment (a), respectively. These scores are identical for predictor groups who submitted models for all 57 hAUs.

The sum of model Z scores for each metric by each group (Tables I and andII)II) provides a useful assessment of performance with respect to that particular metric. However, for ranking the performance of the predictor groups it is useful to have a single composite performance score. For CASP10, the composite performance score (Sums in Table I and Table II) was computed by averaging, for each predictor group, the summed Z scores for the four metrics.

An average performance score per AU was then calculated by dividing this composite score by the number of AUs. This average performance score for each group was computed in two ways: (i) dividing the composite performance score by the total number of available AUs (Avg-a) (i.e. 112 or 57), or (ii) dividing by the total number of AUs submitted by the group (Avg-s). The former method penalizes groups that did not submit predictions for all AUs. For the top performing groups, all of which provided predictions for 111–112 hsAUs (Table I) [or for 56–57 of the hAUs (Table II)], these average values were approximately the same. The rankings shown in Tables I and andIIII are based on the sum average Avg-a, which slightly penalizes the groups that did not submit predictions for all of the available AUs.

The sum MolProbity21,22 Z scores for each predictor group is also tabulated for the top 25 performing groups in Tables I and andIIII (and in Supplementary Tables I and II for all groups). Although these MolProbity scores were not used in ranking predictor groups, they provide a knowledge-based assessment of structure quality. In particular, the Baker group consistently provided models with good MolProbity scores.

Head-to-head Paired Student’s t-test

The Z score ranking was not used to provide a final ranking of predictor groups. As in CASP91, the Z score ranking (based on the sum average Avg-a scores) was only used to identify the top 25 groups, which were then selected for a more detailed assessment accounting for the statistical significance of ranking one group above another. The predictions of these 25 groups were then compared in a direct head-to-head statistical analysis on common targets. In this analysis, each score distribution (GDT-HA, GDC-all, LDDT-15, RPF-9) was considered separately. Using each of these metrics independently, the raw score distributions for each of the AUs from each predictor group were compared with the corresponding score distributions from other groups in the top 25 list, using pair-wise Student’s t-test. The number of comparisons for which there was a statistically significant difference was then summed for each metric. For each score metric, each of the top 25 groups is compared with 24 other groups. Group A was assigned 1 point if its distribution of raw scores for that metric is significantly better (p < 0.05) than that of group B. Hence, the maximum score per metric is 24; and the maximum summed score for four metrics assessed is 96. For models with Z scores < −2, the raw score was set to the raw score value for that AU corresponding to Z = −2 (based on the second cycle Z score analyses summarized in Tables I and andII).II). This consideration was relatively insignificant for most of the 25 top-performing groups, for which even the worst models generally have Z > −2. When a model was missing for a group, the median score for that measure was used as the raw score. This consideration was also insignificant, as most of the 25 top-performing groups submitted prediction models for all AU targets. These summed Student’s t-test head-to-head pairwise scores (H2H Scores), plotted in Figure 4A (for the 112 hsAUs) and Figure 4B (for the 57 hAUs), provided the basis for determining the final ranking of predictor group.

An external file that holds a picture, illustration, etc.
Object name is nihms548634f4.jpg
Ranking of top 25 server-only and human / server predictor groups

(A) Head-to-head pairwise Student’s t-test analysis on raw scores between 25 top-ranking server predictor groups for 112 hsAUs. (B) Head-to-head pairwise Student’s t-test analysis on raw scores between 25 top-ranking server / human predictor groups for 57 hAUs. Black, GDT-HA; red, GDC-all; green, RPF-9; blue, LDDT-15. Top-ranking 25 groups were identified based on average Z score Avg-a (Tables I and andIIII).

It is important to recognize that the final ranking among the top 25 performing groups did not depend on the details of our methods for computing Z scores; the Z scores were used only to provide an overall ranking of all groups (presented in Supplementary Tables I and II), and to identify the top 25 performing groups. The ranking within these top 25 groups was based on the raw scores for each of the four metrics, using paired Student’s t-test in head-to-head comparisons of the predictions made by each group on the accuracy of the predictions of the same target by all of the other 24 groups. In all of the H2H analyses presented in this paper, the order of the groups shown in the H2H analysis plots for which the distribution of raw prediction scores for each of the four metrics assessed were not statistically better than any of the other top 24 groups, are random. The relative rankings of these groups are determined by the Z score statistics of Supplementary Tables S-I, S-II, and S-III.

H2H Ranking – Servers

Figure 4A summarizes the results of the head-to-head (H2H) Student’s t-test analysis for the top 25 performing server groups, assessed on 112 AUs. The five top performing groups were Zhang Server (035s), QUARK (114s), PMS (108s), HHpred-thread (370s), and RaptorX-ZY(122s). These were followed by BAKER-ROSETTASERVER (330s), HHPredA (430s), RaptorX (486s), HHPredAQ (223s), and MULTICOM-NOVEL (424s).

Generally, the H2H ranking of Figure 4A correlates with the Z score ranking of Table I: e.g. 330s > 035s > 108s > 114s > 370s > 122s > 430s > 223s > 486s > 424s. One notable outlier in this correlation is the BAKER-ROSETTASERVER (330s), which was top-ranked in the Z score analysis (Table I), but sixth-ranked in the H2H ranking (Figure 4A). Figure 5 compares GDT-HA raw score and Z score distributions for predictions made by Zhang Server (035s) and BAKER-ROSETTASERVER (330s). This analysis demonstrates that, based on raw GDT-HA scores (Figure 5A), 035s had better performance on the generally more challenging hAUs (i.e. targets with low sequence identity with templates) than 330s. Performance by 330s was marginally better on soAUs, with easily identified templates of known structure. These results suggest that 035s was able to do a superior job than 330s in identifying distant templates. Although this same better performance on more challenging hAUs is evident in the Z score analysis (Figure 4B), these Z scores are much less discriminating than the raw scores, particularly for the challenging hAUs with very low sequence identity with templates. The classification of soAUs and hAUs was made by the CASP Prediction Center based on the HHSearch probability score with potential templates, and may not always accurately reflect the actual difficulty of the targets. In addition, we observed many examples where Z scores overestimate prediction performance when the spreads of the metric for some specific targets are narrow (Supplementary Figure S6). Overall, these results demonstrate the value of using the raw accuracy metric scores, through the H2H paired Student’s t-test, in determining the final ranking.

An external file that holds a picture, illustration, etc.
Object name is nihms548634f5.jpg
Raw metric scores provide more sensitivity to poor performance on more challenging hAUs

Comparison of performance by top-ranking groups shows how (A) GDT-HA raw scores are more sensitive to poor performance on more challenging hAUs than (B) GDT-HA Z scores. Close circles, hAU targets. Open circles, soAU targets. The three dashed lines of Panel A are at y = × and y = × ± 20 units, and the three dashed lines of Panel B are at y = × and y = × ± 1.5 units. Panel B is expanded to exclude regions of the plot with no data points.

We also explored the sensitivity of the H2H ranking against some extremely poor predictions. One approach for suppressing the high impact of those extreme outliers is to replace the Student’s t-test with the non-parametric Wilcoxon signed rank test. This statistic compares the rank, rather than value, of the raw score differences between two distributions, and is less sensitive to raw score outliers. As illustrated in Supplementary Figure S2, suppression of outliers using the pairwise Wilcoxon rank sum analysis alters the ranking among the top 25 performing groups; e.g. in this analysis BAKER-ROSETTASERVER (330s) becomes the top ranked server.

H2H Ranking – Human and/or Servers

Figure 4B summarizes the results of the head-to-head (H2H) paired Student’s t-test analysis for the top 25 performing human and/or server groups, assessed on the 57 hAUs. The score for the top-performing group, Zhang (237), was significantly higher than any other group. The next best performing groups were Leecon (027) and Zhang-Server (035s), followed by TASSER (079), Pcomb (130), Pcons (267), QUARK (114s), MULTICOM (489), CNIO (475), and Mufold (197).

For the Human and/or Server analysis, the H2H ranking was generally well-correlated with the Z score ranking (Table II); 237 > 027 > 035s > 130 > 197 > 79 > 267 > 489 > 344 > 114s. The differences between H2H and Z score ranking again could be traced to the higher sensitivity of raw score comparisons over Z score comparison, especially for hAUs with low sequence identity with templates.

H2H Ranking – Most Difficult TBMs

We also ranked the human and/or server predictors using only the 15 “most difficult” AUs. Figure 6 is a bar plot of the maximum GDT-TS (Max GDT-TS) score obtained for each of the 112 hsAU’s by any of the predictors. AUs with Max GDT-TS ≤ 50% are referred to here as TBM_hard AUs. Based on this analysis, 15 CASP10 AUs are defined as the “TBM_hard AUs”: T0649_D1, T0653_D1, T0668_D1, T0671_D2, T0676_D1, T0678_D1, T0684_D1, T0690_D1, T0705_D2, T0717_D2, T0726_D1, T0732_D2, T0735_D1, T0739_D3, T0739_D4. Using only these 15 targets, we repeated the Z score analysis, shown in Table III for the top-ranked 25 groups, and in Supplementary Table S-III for all groups. We also carried out head-to-head paired Student’s t-test analysis of raw scores among the top-performing 25 groups. Figure 7 summarizes the results of the H2H, assessed on only the 15 TBM_hard AUs. For this set of targets, the top-performing groups were zhang (237) and MULTICOM (489), with nearly identical performance that was significantly better than that for the other 23 groups. These were followed by MuFold (197), TASSER (079), Pcons (267), chuo-fams (365), RaptorX-ZY (122s), Pcomb (130), Kloczkowski (350) and Leecon (027). Significantly, of the 69 server predictors participating in CASP10, only one server, RaptorX-ZY (122s), performed well with these TBM_hard AUs.

An external file that holds a picture, illustration, etc.
Object name is nihms548634f6.jpg
Classifying TBM_hard AUs

Bar plot of the maximum GDT-TS (Max GDT-TS) score obtained for each of the 112 AU’s (indicated by AU target id) by any of the predictors. The 15 AU targets with Max GDT-TS ≤ 50% were identified as TBM_hard AUs.

An external file that holds a picture, illustration, etc.
Object name is nihms548634f7.jpg
Ranking of top 25 human / server predictor groups on TBM_hard targets

Head-to-head (H2H) pairwise Student’s t-test analysis on raw scores between 25 top-ranking server/human predictor groups for 15 TBM_hard AUs. Black, GDT-HA; red, GDC-all; green, RPF-9; blue, LDDT-15. Top-ranking 25 groups were identified based the Avg-a Z score.

Table III

Sum and Average Z-scores for Top 25 Performing Predictor Groups – 15 TBM_Hard AUs.

GroupNameN_modelGDT-HAGDC-allRPFLDDTSumAvg-aAvg-sMolPro
489MULTICOM157.246.5814.7611.9910.140.680.681.48
237zhang156.844.9512.4610.988.810.590.59−0.84
197Mufold146.023.9212.0910.408.110.540.58−6.17
267Pcons155.284.3911.3710.377.850.520.521.74
428PconsQ154.914.0410.909.807.410.490.49−0.36
365chuo-fams154.682.8811.7710.177.380.490.491.99
27LEEcon155.083.3810.559.447.110.470.473.17
130Pcomb154.553.0511.109.176.970.460.468.60
035sZhang-Server155.183.559.858.886.870.460.46−3.49
344Jones-UCL144.303.4710.258.546.640.440.47−13.90
294chuo-repack153.432.4410.629.406.470.430.43−3.02
350Kloczkowski_Lab153.182.3710.849.006.350.420.42−1.04
285McGuffin144.133.159.137.776.050.400.43−0.45
475CNIO153.551.799.998.625.990.400.40−1.58
79TASSER155.566.947.173.985.910.390.39−2.92
434chuo-fams-consensus152.831.559.518.675.640.380.38−0.49
114sQUARK143.832.738.457.315.580.370.40−5.13
315keasar153.882.249.026.395.380.360.36−15.88
122sRaptorX-ZY155.885.625.563.755.200.350.35−14.48
45Zhang_Ab_Initio152.963.377.686.715.180.350.351.19
490Zhang_Refinement154.002.567.526.405.120.340.344.24
26Pr oQ2clust142.783.416.796.684.920.330.35−1.75
101WeFold142.942.388.165.634.780.320.34−9.59
388ProQ2152.111.118.257.484.740.320.3220.29
330sBAKER-ROSETTASERVER152.632.577.265.784.560.300.3029.53

The columns labeled GDT-HA, GDC-all, RPF, and LDDT are the sum of Z scores across all models submitted by each predictor group. The Sum column is the average of the sum of Z scores for the four metrics assessed. The Avg-s and Avg-a scores are the Sum scored divided by the number of AUs for which a model was submitted (s) by each predictor group, and the total number of AUs used for assessment (a), respectively. These scores are identical for predictor groups who submitted models for all TBM_Hard hAUs.

It was particularly important to use multiple metrics of model accuracy in assessing performance on the TBM_hard AUs, as some groups scored well with some metrics, other groups scored well with different metrics. Only the two top-ranked groups zhang (237) and MULTICOM (489) consistently scored well with all four metrics. This is distinct from the results with 112 hsAUs or 57 hAUs, in which there was generally consistent performance on all four metrics by the top scoring groups.

Impact of Distance Matrix Methods in Assessing Low Accuracy Structures

Figure 7 shows that the superimposition-independent LDDT-15 and RPF-9 scores are helpful in distinguishing performance of predictor groups on the TBM_hard AU’s. Also relevant was the complementary value of LDDT-15 and RPF-9 scores in ranking TBM_hard AUs; some predictor groups did best with LDDT-15 scores, while other did best with RFP-9 scores.

Unlike other metrics used in CASP10, the RPF-9 score is normalized against a free rotating chain model, based on Flory polymer chain statistics25,28. For this reason, RPF-9 DP scores used in CASP10 are very discriminative against random structures. Random-like incorrect prediction models (i.e. low quality structures) will have RPF-9 scores very close to zero. Structures with incorrect secondary structures or incorrect folds can even have negative RPF-9 scores, indicating that they are even worse than random structures. Examples discussed in the Supplemental Material demonstrate that the RPF-9 scores normalized to random structure (i.e. the RPF-9 DP scores) have stronger discriminating power than the GDT_HA scores against structures with random-like incorrect folds (Supplementary Figure S3). Normalization against a set of well-defined decoy models, representing a random distribution of structures, has also recently been introduced into the LDDT score24. Additional examples comparing the sensitivity of global alignment (GDT and GDC-all) scores and superimposition-independent (LDDT and RPF) scores for assessing surface loop and interhelical packing inaccuracies are also presented in the Supplementary Material (Supplementary Figure S4).

Figure 7 also shows that TASSER (079), RaptorX-ZY (122s), and MULTICOM (489) had the best performance on TBM_hard AUs with the GDC-all score, although the first two of these groups did not do so well with other three scores. 3rd ranked Mufold (197) had a zero GDC-all count. It appears that TASSER and RaptorX-ZY may have some part of the core packing matched with some experimental structures, yielding higher GDC-all scores even though the main chains of these models may not be more accurate than other models from other groups. However, global alignment scores like GDC-all are challenged when the overall backbone structure is inaccurate, suggesting that GDT-HA and GDC-all scores may be unreliable for assessing predictions of difficult TBM-hard AUs. Despite some of the valuable features of the superimposition-independent metrics like RPF and LDDT for difficult targets like TBM-hard AUs, human judgment like that used in the CASP10 FM assessment14 is still a better approach than using automated scoring metrics.

Impact of accurate model selection

As discussed above in the H2H assessment of severs, a few poor models can make a significant impact on the relative ranking, particularly in distinguishing the top-performing group from other groups that generally did very well in CASP10. In many cases, predictor groups included in their five submitted models more accurate models than the one designated as Model 1. This is a well-recognized aspect of previous CASP experiments. Interestingly, our analysis suggests that at least in some cases, the more accurate model could have been selected using alternative fold-accuracy discriminators. Figure 8 illustrates two examples, one for server BAKER-ROSETTASERVER (330s) and one for human-curated group Sternberg (458) where an alternative model is more accurate (higher GDT-HA raw score) than the designated Model #1. In these two cases, the more accurate model also has a significantly better ProsaII23 fold score.

An external file that holds a picture, illustration, etc.
Object name is nihms548634f8.jpg
The challenge of accurate model selection

Predictor groups often submitted a much less accurate model as Model #1 compared to other of their own submitted alternative models. (A) For AU target T0742_D1, Model #1 submitted by group 035s (Zhang-Server) has much higher GDT-HA score (∼ 0.3; y-axis) than Model #1 submitted by server group 330s (BAKER-ROSETTASERVER; GDT-HA ∼ 0.05). However, alternative Model #3 of group 330s (GDT-HA ∼0.28) is a more accurate model. (B) For AU target T0726_D2, Model #1 submitted by the human-curated group 237 (zhang) group had much higher GDT-HA score (∼ 0.72) than Model #1 submitted by group 458 (Sternberg; GDT-HA ∼ 0.12)). However, alternative Model #2 of group 458 (GDT-HA ∼ 0.71) is a much more accurate model. In these particular cases, the Prosa Z score (x-axis) could have provided a good criteria in model selecting, as it would have indicated a much more accurate “best model” for groups 330s and 458.

We also did a comparison of GDT-HA scores between models picked from among submitted models by ProsaII score only, and the Model #1 provided by predictors for all 10,902 combinations of Groups and AUs for which ProsaII scores are available from the Prediction Center. If all the groups used ProsaII to pick their best model from the five submitted, 91 groups would improve their average GDT-HA score. Our point is not that ProsaII is the single best indicator of model quality. However, as is generally appreciated, many CASP10 participants would benefit by more successful model selection.

Ranking Based on Ideal Model Selection

Selection of the most-accurate model among submitted alternate models is an important area for development in order to provide more accurate TBM predictions. These trends were validated by simulating an “ideal model selection” CASP10 competition, choosing for each group and AU the single model with highest GDT-HA as Model #1, and repeating the Z score and H2H raw score rankings as outlined in the previous sections. Rankings based on H2H Student’s t-test raw score comparisons are shown in Figure 9A for server predictors and Figure 9B for human-curated predictors. While “ideal model selection” has modest impact on server rankings, it has a significant impact on human-curator predictor rankings (cf. the ranking of predictors in Figures 4 and and9).9). Server predictors for which ranking was significantly improved in this simulated ideal model selection CASP10 competition included BAKER-ROSETTASERVER (330s), PconsM (103s), MULTICOM-CLUSTER (081S), and MULTICOM-REFINE (125s). Human-curated predictors for which ranking was significantly improved by simulated ideal model selection include CNIO (475) (which using ideal model selection ranked second behind Zhang (237), Baker (477), chuo-repack (294), chuo-fams (365), Zhang-refinement (490), and keasar (315). These results demonstrate the key role of model selection in accurate TBM protein structure prediction.

An external file that holds a picture, illustration, etc.
Object name is nihms548634f9.jpg
Ideal model selection

For each predictor group, the submitted model with highest GDT-HA score was selected and the paired Student’s t-test analysis was done as in Fig. 4. (A) Head-to-head pairwise Student’s t-test analysis on raw scores between 25 top-ranking server predictor groups for 112 hsAUs. (B) Head-to-head pairwise Student’s t-test analysis on raw scores between 25 top-ranking server / human predictor groups for 57 hAUs. Black, GDT-HA; red, GDC-all; green, RPF-9; blue, LDDT-15. Top-ranking 25 groups were identified based the Avg-a Z score.

Conclusions

The CASP10 TBM assessment identified the Zhang-Server (035s), QUARK (114s), PMS (108s), Leecon (027), Zhang (237) groups as providing the most accurate models for the AU targets. In the course of our assessment, we observed many examples where raw scores were more discriminating for distinguishing structural accuracy than Z scores (also see Supplementary Figure S6), or where Z scores suggest good prediction performance while the raw scores together with visual inspecting reveal inaccurate models (e.g. Figure 3). In our experience, rankings based on raw scores, using head-to-head (H2H) comparisons of performance on common AUs, with a statistical assessment of whether one group outperforms each of the other groups, provides a more sensitive automated numerical assessment than Z scores alone. The H2H ranking developed by Mariani et al1 incorporates Student’s t-test, providing as statistically sound basis for ranking the relative performance of groups on the same set of targets.

The CASP10 TBM assessment, evolving from the previous CASP assessment protocols, put heavy emphasis on side-chain atom positions, which are assessed by three (GDC-all, LDDT-15, and RPF-9) of the four structure accuracy metrics that contribute to our composite score. The philosophy was to assess TBM models with the same kinds of expectations for physically-reasonable structures as is expected for experimental NMR or X-ray crystal structures. We also combined both global alignment scores (GDT-HA and GDC-all) which utilized superimpositions, and distance matrix based methods (LDDT-15 and RPF-9) which do not involve superimposition methods and are more sensitive to local structure accuracy and core side-chain packing features. The distance matrix based methods seem to be less sensitive to the loosely packed loop regions and also to helix tilt angles. The introduction of normalizing the RPF score against a random coil model impacts the ranking for the difficult targets. These multiple scores incorporate information on all heavy atom (C, N, O, and S) positions, and balance the biases inherent in superimposition-based structural comparisons (e.g. GDT scores) which are problematic when comparing structures lacking high overall structural similarity.

Supplementary Material

Supplementary Material

Supplementary Material

Click here to view.(3.3M, pdf)

Acknowledgments

We thank all the scientists who contributed experimental structures to the CASP10 project, without which this project would not have been possible. Special thanks to A. Kryshtafovych for extensive efforts to support our study by providing guidance based on historical CASP assessment projects, for implementing new programs on the CASP Prediction Center cpu cluster, and for running the extensive numerical analyses that were the basis for our assessment. We also thank B. K. Lee, C.-H. Tai, H. Bai, J. Block, K. Fidelis, J. Moult, T. Schwede, T. J. Taylor, and A. Tramontano for extensive scientific discussions and helpful criticisms. This work was supported by a grant from the Protein Structure Initiative of the National Institutes of Health (U54-{"type":"entrez-nucleotide","attrs":{"text":"GM094597","term_id":"221870710","term_text":"GM094597"}}GM094597).

Center for Advanced Biotechnology and Medicine, Rutgers, The State University of New Jersey and Robert Wood Johnson Medical School, University of Medicine and Dentistry of New Jersey, and Northeast Structural Genomics Consortium, 679 Hoes Lane, Piscataway, New Jersey, 08854, USA
To whom correspondence should be addressed: Prof. Gaetano T. Montelione CABM, Rutgers University, 679 Hoes Lane, Piscataway, NJ 08854-5638, Phone: 732-235-5321, ude.sregtur.mbac@yug
YJH and BM contributed equally to this study and are designated as co-first authors.

Abstract

Template Based Modeling (TBM) is a major component of the Critical Assessment of Protein Structure Prediction (CASP). In CASP10, some 41,740 predicted models submitted by 150 predictor groups were assessed as TBM predictions. The accuracy of protein structure prediction was assessed by geometric comparison with experimental X-ray crystal and NMR structures using a composite score that included both global alignment metrics and distance-matrix based metrics. These included GDT-HA and GDC-all global alignment scores, and the superimposition-independent LDDT distance-matrix based score. In addition, a superimposition-independent RPF metric, similar to that described previously for comparing protein models against experimental NMR data, was used for comparing predicted protein structure models against experimental protein structures. In order to score well on all four of these metrics, models must feature accurate predictions of both backbone and side-chain conformations. Performance rankings were determined independently for server and the combined server plus human-curated predictor groups. Final rankings were made using pair-wise head-to-head Student’s t-test analysis of raw metric scores among the top 25 performing groups in each category.

Keywords: CASP10, protein structure prediction, GDT score, LDDT score, RPF DP scores, structural bioinformatics, homology modeling, comparative modeling
Abstract

Abbreviations

AUCASP10 assessment units, corresponding to regions of experimental structures used in assessing model predictions. FM, free modeling
TBMtemplate-based modeling
TBM hsAUsall 112 TBM AUs, including AUs for which models were provided by either server or human-curated predictors
TBM soAUs55 AUs which for which only predictions by servers were provided and assessed
TBM hAUsthe subset of 57 hsAUs for which human-curated predictions were provided and assessed
Abbreviations

References

  • 1. Mariani V, Kiefer F, Schmidt T, Haas J, Schwede TAssessment of template based protein structure predictions in CASP9. Proteins. 2011;79(Suppl 10):37–58.[PubMed][Google Scholar]
  • 2. Schwede T, Sali A, Honig B, Levitt M, Berman HM, Jones D, Brenner SE, Burley SK, Das R, Dokholyan NV, Dunbrack RL, Jr, Fidelis K, Fiser A, Godzik A, Huang YJ, Humblet C, Jacobson MP, Joachimiak A, Krystek SR, Jr, Kortemme T, Kryshtafovych A, Montelione GT, Moult J, Murray D, Sanchez R, Sosnick TR, Standley DM, Stouch T, Vajda S, Vasquez M, Westbrook JD, Wilson IAOutcome of a workshop on applications of protein models in biomedical research. Structure. 2009;17(2):151–159.[Google Scholar]
  • 3. Moult JComparative modeling in structural genomics. Structure. 2008;16:14–16.[PubMed][Google Scholar]
  • 4. Liu J, Montelione GT, Rost BNovel leverage of structural genomics. Nature biotechnology. 2007;25(8):849–851.[PubMed][Google Scholar]
  • 5. Nair R, Liu J, Soong TT, Acton TB, Everett JK, Kouranov A, Fiser A, Godzik A, Jaroszewski L, Orengo C, Montelione GT, Rost BStructural genomics is the largest contributor of novel structural leverage. Journal of structural and functional genomics. 2009;10(2):181–191.[Google Scholar]
  • 6. Montelione GTThe Protein Structure Initiative: achievements and visions for the future. F1000 Biol Rep. 2012;4:7.[Google Scholar]
  • 7. Zhang Y, Thiele I, Weekes D, Li Z, Jaroszewski L, Ginalski K, Deacon AM, Wooley J, Lesley SA, Wilson IA, Palsson B, Osterman A, Godzik AThree-dimensional structural view of the central metabolic network of Thermotoga maritima. Science. 2009;325(5947):1544–1549.[Google Scholar]
  • 8. Zhang QC, Petrey D, Deng L, Qiang L, Shi Y, Thu CA, Bisikirska B, Lefebvre C, Accili D, Hunter T, Maniatis T, Califano A, Honig BStructure-based prediction of protein-protein interactions on a genome-wide scale. Nature. 2012;490(7421):556–560.[Google Scholar]
  • 9. Huang YJ, Hang D, Lu LJ, Tong L, Gerstein MB, Montelione GTTargeting the human cancer pathway protein interaction network by structural genomics. Molecular &amp; cellular proteomics : MCP. 2008;7(10):2048–2060.[Google Scholar]
  • 10. Qian B, Raman S, Das R, Bradley P, McCoy AJ, Read RJ, Baker DHigh-resolution structure prediction and the crystallographic phase problem. Nature. 2007;450(7167):259–264.[Google Scholar]
  • 11. Raimondo D, Giorgetti A, Bosi S, Tramontano AAutomatic procedure for using models of proteins in molecular replacement. Proteins. 2007;66(3):689–696.[PubMed][Google Scholar]
  • 12. Terwilliger TC, Dimaio F, Read RJ, Baker D, Bunkoczi G, Adams PD, Grosse-Kunstleve RW, Afonine PV, Echols Nphenix.mr_rosetta: molecular replacement and model rebuilding with Phenix and Rosetta. Journal of structural and functional genomics. 2012;13(2):81–90.[Google Scholar]
  • 13. Keedy DA, Williams CJ, Headd JJ, Arendall WB, 3rd, Chen VB, Kapral GJ, Gillespie RA, Block JN, Zemla A, Richardson DC, Richardson JSThe other 90% of the protein: assessment beyond the Calphas for CASP8 template-based and high-accuracy models. Proteins. 2009;77(Suppl 9):29–49.[Google Scholar]
  • 14. Tai C-H, Bai H, Taylor TJ, Lee BKAssessment of template free modeling in CASP10 and ROLL. Proteins. 2013 Accompanying paper. [[PubMed][Google Scholar]
  • 15. Cozzetto D, Kryshtafovych A, Fidelis K, Moult J, Rost B, Tramontano AEvaluation of template-based models in CASP8 with standard measures. Proteins. 2009;77(Suppl 9):18–28.[Google Scholar]
  • 16. Kryshtafovych A, Fidelis K, Tramontano AEvaluation of model quality predictions in CASP9. Proteins. 2011;79(Suppl 10):91–106.[Google Scholar]
  • 17. Taylor TJ, Tai C-H, Huang YJ, Block J, Bai H, Kryshtafovych A, Montelione GT, Lee BKDefinition and classifcation of evaluation units for CASP10. Proteins. 2013 accompanying paper. [Google Scholar]
  • 18. Snyder DA, Grullon J, Huang YJ, Tejero R, Montelione GTThe Expanded FindCore method for identification of a core atom set for assessment of protein structure prediction. Proteins. 2013 Accompanying paper. [Google Scholar]
  • 19. Kryshtafovych A, Monastyrskyy B, Fidelis KCASP Prediction Center infrastructure and evaluation measures in CASP10 and CASP ROLL. Proteins. 2013 Accompanying paper. [Google Scholar]
  • 20. Zemla ALGA: A method for finding 3D similarities in protein structures. Nucleic Acids Res. 2003;31(13):3370–3374.[Google Scholar]
  • 21. Chen VB, Arendall WB, 3rd, Headd JJ, Keedy DA, Immormino RM, Kapral GJ, Murray LW, Richardson JS, Richardson DCMolProbity: all-atom structure validation for macromolecular crystallography. Acta crystallographica Section D, Biological crystallography. 2010;66(Pt 1):12–21.[Google Scholar]
  • 22. Lovell SC, Davis IW, Arendall WB, 3rd, de Bakker PI, Word JM, Prisant MG, Richardson JS, Richardson DCStructure validation by Calpha geometry: phi,psi and Cbeta deviation. Proteins. 2003;50(3):437–450.[PubMed][Google Scholar]
  • 23. Sippl MJRecognition of errors in three-dimensional structures of proteins. Proteins. 1993;17(4):355–362.[PubMed][Google Scholar]
  • 24. Mariani V, Biasini M, Barbato A, Schwede TlDDT: a local superposition-free score for comparing protein structures and models using distance difference tests. Bioinformatics. 2013[Google Scholar]
  • 25. Huang YJ, Powers R, Montelione GTProtein NMR recall, precision, and F-measure scores (RPF scores): structure quality assessment measures based on information retrieval statistics. J Am Chem Soc. 2005;127(6):1665–1674.[PubMed][Google Scholar]
  • 26. Rosato A, Aramini JM, Arrowsmith C, Bagaria A, Baker D, Cavalli A, Doreleijers JF, Eletsky A, Giachetti A, Guerry P, Gutmanas A, Guntert P, He Y, Herrmann T, Huang YJ, Jaravine V, Jonker HR, Kennedy MA, Lange OF, Liu G, Malliavin TE, Mani R, Mao B, Montelione GT, Nilges M, Rossi P, van der Schot G, Schwalbe H, Szyperski TA, Vendruscolo M, Vernon R, Vranken WF, de Vries S, Vuister GW, Wu B, Yang Y, Bonvin AMBlind testing of routine, fully automated determination of protein structures from NMR data. Structure. 2012;20(2):227–236.[Google Scholar]
  • 27. Huang YJ, Rosato A, Singh G, Montelione GTRPF: a quality assessment tool for protein NMR structures. Nucleic Acids Res. 2012;40(Web Server issue):W542–W546.[Google Scholar]
  • 28. Flory PJ Statistical mechanics of chain molecules. New York: Interscience Publishers; 1969. [PubMed][Google Scholar]
  • 29. Read RJ, Chavali GAssessment of CASP7 predictions in the high accuracy template-based modeling category. Proteins. 2007;69(Suppl 8):27–37.[PubMed][Google Scholar]
  • 30. Kopp J, Bordoli L, Battey JN, Kiefer F, Schwede TAssessment of CASP7 predictions for template-based modeling targets. Proteins. 2007;69(Suppl 8):38–56.[PubMed][Google Scholar]
  • 31. Clarke ND, Ezkurdia I, Kopp J, Read RJ, Schwede T, Tress MDomain definition and target classification for CASP7. Proteins. 2007;69(Suppl 8):10–18.[PubMed][Google Scholar]
  • 32. Tress ML, Ezkurdia I, Richardson JSTarget domain definition and classification in CASP8. Proteins. 2009;77(Suppl 9):10–17.[Google Scholar]
  • 33. Kinch LN, Shi S, Cheng H, Cong Q, Pei J, Mariani V, Schwede T, Grishin NVCASP9 target classification. Proteins. 2011;79(Suppl 10):21–36.[Google Scholar]
  • 34. Soding J, Biegert A, Lupas ANThe HHpred interactive server for protein homology detection and structure prediction. Nucleic Acids Res. 2005;33(Web Server issue):W244–W248.[Google Scholar]
  • 35. Kryshtafovych A, Krysko O, Daniluk P, Dmytriv Z, Fidelis KProtein structure prediction center in CASP8. Proteins. 2009;77(Suppl 9):5–9.[Google Scholar]
  • 36. Cozzetto D, Kryshtafovych A, Tramontano AEvaluation of CASP8 model quality predictions. Proteins. 2009;77(Suppl 9):157–166.[PubMed][Google Scholar]
  • 37. Tramontano A, Morea VAssessment of homology-based predictions in CASP5. Proteins. 2003;53(Suppl 6):352–368.[PubMed][Google Scholar]
Collaboration tool especially designed for Life Science professionals.Drag-and-drop any entity to your messages.