 Methodology
 Open Access
 Published:
SOV_refine: A further refined definition of segment overlap score and its significance for protein structure similarity
Source Code for Biology and Medicinevolume 13, Article number: 1 (2018)
Abstract
Background
The segment overlap score (SOV) has been used to evaluate the predicted protein secondary structures, a sequence composed of helix (H), strand (E), and coil (C), by comparing it with the native or reference secondary structures, another sequence of H, E, and C. SOV’s advantage is that it can consider the size of continuous overlapping segments and assign extra allowance to longer continuous overlapping segments instead of only judging from the percentage of overlapping individual positions as Q3 score does. However, we have found a drawback from its previous definition, that is, it cannot ensure increasing allowance assignment when more residues in a segment are further predicted accurately.
Results
A new way of assigning allowance has been designed, which keeps all the advantages of the previous SOV score definitions and ensures that the amount of allowance assigned is incremental when more elements in a segment are predicted accurately. Furthermore, our improved SOV has achieved a higher correlation with the quality of protein models measured by GDTTS score and TMscore, indicating its better abilities to evaluate tertiary structure quality at the secondary structure level. We analyzed the statistical significance of SOV scores and found the threshold values for distinguishing two protein structures (SOV_refine > 0.19) and indicating whether two proteins are under the same CATH fold (SOV_refine > 0.94 and > 0.90 for three and eightstate secondary structures respectively). We provided another two example applications, which are when used as a machine learning feature for protein model quality assessment and comparing different definitions of topologically associating domains. We proved that our newly defined SOV score resulted in better performance.
Conclusions
The SOV score can be widely used in bioinformatics research and other fields that need to compare two sequences of letters in which continuous segments have important meanings. We also generalized the previous SOV definitions so that it can work for sequences composed of more than three states (e.g., it can work for the eightstate definition of protein secondary structures). A standalone software package has been implemented in Perl with source code released. The software can be downloaded from http://dna.cs.miami.edu/SOV/.
Background
Protein secondary structure (SS) in three (H for helix, E for strand, and C for coil) or eight states as defined in [1] is a typical example of segmented sequences in bioinformatics. Besides protein secondary structure, new bioinformatics problems arose recently that were also dealing with segmented sequences. For example, topologically associating domains (TADs) were recently identified as megabasesized selfinteraction regions in mammalian genomes [2]. Given a genomic region containing several TADs, we can label the bodies of TADs as “D” and the boundary regions as “B”, resulting in a segmented sequence in two states (i.e., D and B). These cases about segmented sequences raise an issue about how to benchmark the predicted sequence against the reference one (e.g., the observed secondary structures), because evaluation methods based on individual positions, such as the Q3 score (that is equal to the ratio between the count of identical positions and the length of sequence), cannot take the length of continuous segments into consideration. Therefore, a measurement that can address this issue is in demand.
Segment overlap measure (SOV) was originally defined in [3] by Rost et al. to evaluate the quality of predicted protein secondary structures on a segment base. It takes several factors into consideration including the number of segments in a secondary structure, the averaged segment length, and the distribution of the length values. As a result, it allows some variations at the boundary regions of the segments by assigning some allowance (bonus), and can handle extreme cases (e.g., penalizing wrong predictions) reasonably by providing a sliding scale of segment overlap. However, that measure did not normalize the SOV scores into a fixed range, which makes it difficult to compare with other scores in percentage terms or in the range of 0 to 1. This problem was addressed by Zemla et al. in [4]. The modified definition of SOV (SOV’99) uses the length of all segments in the reference sequence to normalize the SOV scores for each state, which can make the final SOV score in percentage scale. Both measures define allowance (bonus) mechanisms for allowing some variations at the boundaries of segments, which are very important and can directly affect the scale of SOV scores for each state. For both of these two SOV definitions, the allowance assigned to each overlapping segment pair is determined by several factors including segment length and overlapping level and is an integer that cannot lead to a more than perfect value of SOV score (i.e., larger or equal to 1 for range 0–1). This 1999 version of SOV score (SOV’99) has been widely used as a standard measure for evaluating protein secondary structure predictions [5,6,7,8,9,10,11,12,13,14]. Currently, it also has been widely used in quality assessment (QA) of protein models as a machine learning feature [15,16,17].
However, our analysis will later show that the definition of allowance in SOV’99 has a significant drawback, that is, it cannot ensure the allowance being incremental when the prediction becomes better. For example, suppose there have been five continuous overlapping correctlypredicted positions (identical between predicted and reference sequence, for example, “HHHHH”), if one more position is accurately predicted, i.e., making it a sixelement overlapping segment, more allowance should be given than the previous fiveposition case. The intuition is that accurately predicting one more position on top of a fiveelement segment (e.g., from “HHHHH” to “HHHHHH”) is more difficult and deserves more bonus points. However, the definition of SOV’99 cannot ensure this intuition. In this research, we further modified SOV’99 by designing a new definition of allowance and named it SOV_refine.
Results
In this section, we first explain the advantage of SOV_refine modified from the definition of SOV’99. Specifically, we use an example to illustrate incremental allowance when one more residue is predicted accurately while keeping other advantages of SOV’99. After that, we show that SOV_refine can better indicate the threedimensional quality of protein models at the secondary structure level. We then provide indepth analysis of statistical significance of Q3 and SOV scores. Finally, we demonstrate two application examples of SOV scores: (1) SOV scores as machine learning features for developing quality assessment tools; (2) Evaluating the similarity of the inferred locations of TADs in mammalian genomes.
Incremental allowance for better predictions
In order to make a direct comparison with SOV’99, here we use the same examples provided in the publication of SOV’99 [4]. As shown in Table 1, “predicted 1” seems to be a bad prediction because it does not have an Hstate segment with a length larger than two. Therefore, although it has a relatively high Q3 value, the two SOV scores (SOV’99 and our SOV_refine) are relatively small (both punish this case to different degrees compared to Q3 score). It can be found that the SOV_refine score is slightly larger than the SOV’99 score in this case. This is because the amount of allowance assigned by SOV’99 in this case is zero [4], whereas our SOV_refine is designed to assign a largerthanzero allowance, in this case to the Hstate segment. Obviously, SOV’99 gives this case a larger punishment than our method SOV_refine (SOV’99 gives a zero allowance and a lower overall score than SOV_refine does). However, it should be noticed that although “predicted_1” is a bad prediction, it is not completely wrong because it does accurately predict some isolated states in a couple of positions (it’s just that the accurately predicted positions are not adjacent to each other). However, SOV’99 assigns a harsh punishment by assigning allowance zero, but our SOV_refine still assigns a small allowance, small enough to show that it is a bad prediction that deserves a SOV_refine score to be much lower than Q3 score (i.e. some punishments compared to Q3 score), but meanwhile not as low as a zero allowance, to indicate “predicted 1” is not completely wrong. This makes our SOV_refine more reasonable because it not only can punish bad predictions compared to Q3 score (our SOV_refine gives a much lower score than Q3 score for this example) but also does not give an extremely low allowance (zero) for the bad predictions such as “predicted 1” as it does accurately predict some isolated states.
The next two predictions (i.e., “predicted 2” and “predicted 3”) have longer Hstate segments, resulting in larger SOV scores. Predictions 4 through 8 are deliberately selected to demonstrate the essential difference between SOV’99 and our SOV_refine when one more element (i.e., Hstate residue in predicted assignments) is further predicted accurately. As expected, the accuracy for Q3 is increased by a fixed value of 0.083. For SOV’99, the scores are irregularly increased by 0.008, 0.174, 0.097, and 0.041, while the scores from SOV_refine are increased by 0.108, 0.111, 0.119, and 0.14, which keep increasing when the predictions are getting better.
The two different observations can be properly explained from the distinct definitions of assigning allowance from SOV’99 and SOV_refine. To be specific, SOV’99 cannot ensure the amount of allowance stably increased, whereas SOV_refine is designed to be capable of handling this case.
We provide another example with a reference sequence composed of four states shown in Table 2 to demonstrate that SOV_refine can assign distinguishable scores by adjusting λ parameter in the definition. There are four predictions, which are getting better from predicted 1 up to predicted 4. SOV’99 and SOV_refine (λ = 1) cannot distinguish which one is better between predicted 3 and predicted 4, whereas SOV_refine with λ equal to 0.5 or equal to 0.1 can conclude that predicted 4 is better than predicted 3. This example indicates that the definition of SOV_refine is more flexible than that of SOV’99. It can be found that a smaller λ will make the SOV_refine algorithm more stringent. Table 2 lists all the SOV_refine scores with different λ values, from which users can pick up the appropriate λ value based on their stringency demands of their specific problems. In our implementation, the default value of λ is 1.
Evaluation of protein tertiary models at the secondary structure level
We downloaded the protein native structures and predicted models of 33 TemplateBased Modeling (TBM) singledomain targets in the Critical Assessment of protein Structure Prediction 11 (CASP11) at http://www.predictioncenter.org/casp11/. The native structures for the 33 singledomain targets are available at CASP official website. For each target, 20 protein models in stage 1 for quality assessment (QA) are chosen as the predicted structures [18], because these models cover the whole range of model accuracy.
We then superimposed the 20 models of each target with their native structure using three different protein structure alignment tools: LGA [19], TMalign [20], and DeepAlign [21], resulting in a set of scores (i.e., GDTTS from LGA, TMscore from TMalign, GDTTS from DeepAlign, and TMscore from DeepAlign) for measuring the quality of predicted protein 3D models from threedimensional superimposing. After that, secondary structures of 33 native structures and their corresponding models were assigned by STRIDE [22] and DSSP [1] in three states (i.e., H, E, and C), respectively. Overall, we obtained four sets of 660 GDTTS or TMscore and 660 pairs of observed and predicted secondary structures, for each pair of which we carried out the comparisons of secondary structures using three measures: Q3 score for threestate secondary structure, SOV’99, and SOV_refine (λ = 1), respectively.
We then explored whether our SOV_refine can better indicate the threedimensional quality of protein models by comparisons at the secondary structure level. The Pearson’s correlation coefficients were calculated between each of the four sets of the threedimensional superimposing scores (GDTTS from LGA and DeepAlign, and TMscore from TMalign and DeepAlign) and the scores of comparing secondary structures using Q3 score, SOV’99, and SOV_refine (see Fig. 1(a) for STRIDE and 1(b) for DSSP): for using STRIDE, SOV_refine (λ = 1) constantly achieves the best performance with r = 0.70, 0.71, 0.70, and 0.73 (pvalue < 10^{− 5}), followed by SOV’99 (r = 0.67, 0.70, 0.67, and 0.72), and Q3 (r = 0.60, 0.68, 0.60, and 0.70); for using DSSP, we can draw the same conclusion. Since the λ parameter in the definition of SOV_refine in Eq. 4 is adjustable, we have tried to check whether different λ values affect the Pearson’s correlation performance. We have tested the λ values in the range of [0.1, 2.4]. The results shown in Fig. 2 indicate that smaller λ values achieve larger Pearson’s correlation coefficients.
Statistical significance of Q3 and SOV_refine scores
Here we aim to address two questions as described in [23]: (1) What is the statistical significance of Q3, SOV’99, and SOV_refine? (2) For a given score what is the probability of two proteins having the same fold?
For the statistical significance of Q3, SOV’99, and SOV_refine, we used Top8000 database [24] including 8000 highresolution qualityfiltered protein chains. After filtering out chains with length larger than 200 or less than 80, we obtained 3420 protein chains, resulting in 5,846,490 protein pairs. For each protein pair, we calculated its Q3, SOV’99, and SOV_refine scores. If two protein chains do not have the same length, the scores were calculated between the smaller chain and a sliding window with length equal to the length of the smaller chain on the larger chain (20residue sliding interval). We finally obtained 14,252,776 scores for Q3, SOV’99, and SOV_refine; and their distribution can be found in Fig. 3(a). The Pvalues (i.e., the probability of having a Q3, SOV’99, or SOV_refine score equal to or larger than a certain value) for a given score were calculated by the same way as in [23] and shown in Fig. 3(b). In general, when Q3 ≤ 0.26, SOV’99 ≤ 0.24, and SOV_refine ≤ 0.19, the probability of finding these scores from our sample is close to 1 (these values were found when the Pvalues start to decrease from 0.95), and then the Pvalues decrease rapidly when the scores are getting larger than these thresholds. These findings mean that in order for Q3, SOV’99, and SOV_refine to distinguish two structures from the secondary structure level, their scores need to be higher than these thresholds.
For the probability of protein pairs having the same fold for a given score, we downloaded the latest CATH database (v4.2) [25]. After filtering out the proteins with length less than 80 or larger than 200, we obtained 18,653 proteins, which were classified into 343 folds. For folds with size larger than 100 we only kept the first 100 proteins. We then used DSSP [1] to assign secondary structures for each protein in three and eight states. For protein pairs with different lengths, we calculated their Q3, SOV’99, and SOV_refine scores as follows: (1) The secondary structure sequence in three and eight states of the smaller protein slides gaplessly (i.e., oneresidue sliding interval) along the bigger protein; (2) The final Q3 and SOV scores of the protein pair are the corresponding maximum values on all the possible sliding positions generated from the first step. For three and eight states, we generated two samples individually (two samples for three states and two samples for eight states), one including scores from the protein pairs in the same folds, and the other including scores from the protein pairs in different folds. The distributions of these scores in three and eight states are shown in Figs. 4 and 5, respectively.
The top three plots in both Figs. 4 and 5 indicate the conditional probability, for example, the red lines indicate P(SOV_refine F), which is the conditional probability of SOV_refine score when the two proteins are in different CATH fold families whereas the blue lines indicate \( P\left({SOV}_{} refine\left\overline{F}\right.\right) \), which is the conditional probability of SOV_refine when the two proteins are in the same CATH fold family.
For a given score of Q3, SOV’99, and SOV_refine, we also calculated its posterior probabilities (bottom three plots in Figs. 4 and 5) that the two proteins have the same or different fold as the way described in [23]. For example, the posterior probability of two proteins to have the same fold given a specific SOV_refine score is represented by P(FSOV_{−}refine), whereas \( P\left(\left.\overline{F}\right{SOV}_{} refine\right) \) for not having the same fold. The results are shown in Fig. 4 for threestate and Fig. 5 for eightstate secondary structures with red lines indicating not having the same fold and blue lines indicating having the same fold. The point when the two lines with different colors intersect is the score threshold, above which we think two proteins are having the same fold.
From Figs. 4 and 5, we can observe that scores in the same folds are slightly larger than those in different folds, but not as noticeable as the observation obtained by using TMscore as in [23]. We can also conclude that for three states when a given score from two random proteins meets Q3 ≤ 0.97 and SOV_refine ≤ 0.94, the two proteins have high probability sharing different folds; for eight states when a given score from two random proteins meets Q8 ≤ 0.95 and SOV_refine ≤ 0.90, the two proteins have high probability sharing different folds. This type of conclusion cannot be drawn for SOV’99 based on the data indicating another advantage of our SOV_refine compared to SOV’99.
Application of SOV_refine for protein quality assessment
We proved that SOV scores, especially SOV_refine, are effective machine learning features for protein quality assessment. We used 85 targets from CASP9 and 67 targets from CASP10 as training data and their real GDT_TS scores as objective values. For each target, we randomly selected 150 protein models. For each model, we extracted 32 features, mostly from [17], as the basic feature set and generated three more feature sets: (1) Basic set plus SOV’99 for predicted and assigned secondary structures; (2) Basic set plus SOV_refine (λ = 0.1) for predicted and assigned secondary structures; (3) Basic set plus SOV_refine (λ = 1) for predicted and assigned secondary structures. We used SCRATCH [26] to obtain the predicted secondary structures. We used Random Forest [27] to train the prediction models.
We blindly tested the performance of the QA models trained from the four feature sets on 75 targets in CASP11 in two stages [18]. The evaluation measures are the same as those in official CASP evaluations [18] including (1) the weighted mean of Pearson’s product moment correlation coefficient (wmPMCC), (2) the average loss (Ave loss), (3) the average GDT_TS deviations (Ave ΔGDT), and (4) the Matthews correlation coefficient (MCC). The blind test results are shown in Table 3. All of the three SOV scores play a positive role in improving the performance; and SOV_refine (λ = 1) performs better than SOV_refine (λ = 0.1) and SOV’99 in terms of most of the evaluation criteria.
SOV_refine scores for measuring similarity of different definitions of topologically associating domains (TADs)
The SOV scores have other important applications in comparing segmented sequences besides protein secondary structure sequences. Here we demonstrate an example of using SOV score to measure the similarity of different definitions (in terms of genomic locations) of topologically associating domains (TADs) in mammalian genomes. For the same genomic region, different TADdetection algorithms may infer different TAD locations [28]. Therefore, SOV score can be used here to measure the similarity of different TAD definitions (i.e., the similarity about which part of the genomic region is within a TAD body and which part is within the boundary).
The normalized HiC data for male mouse embryonic stem cells (mESC) was downloaded from Ren Lab's website at http://chromosome.sdsc.edu/mouse/hic/download.html. As shown in Fig. 6, we selected a genomic region (137.8 Mb – 140.28 Mb) on chromosome 2. There are two TADs in this region based on the definition from Ren Lab http://chromosome.sdsc.edu/mouse/hic/download.html. However, the boundary regions between the two TADs are vague as the HiC signals are not sharply distinguishable. Therefore, different TADdetection algorithms may give different definitions about the locations of the two TADs (or the boundary regions). To illustrate this, we artificially made up the reference definition (based on Ren Lab’s definition) and two other definitions as inference definitions. We labeled “D” for the positions within a TAD body and “B” for positions within a boundary region. In this way, we have three sequences (i.e., one reference and two inferences), each containing two states. The SOV’99 and SOV_refine (λ = 1) between reference and inference 1 (we use “predicted 1” in the figure to match previous examples) are 0.99 and 0.91 respectively. The SOV’99 and SOV_refine (λ = 1) between reference and predicted 2 are 1.0 and 0.89 respectively. It can be found that SOV’99 scores indicate that the two predictions are almost the same as the reference, which is actually not. However, SOV_refine scores can quantitatively detect the differences by giving a lower score, demonstrating another advantage of SOV_refine compared to SOV’99.
Discussion
One may argue that the SOV score which was originally introduced in 1990s and protein secondary structure prediction have already been an old topic and achieved stable performance. Therefore, SOV score may not be needed. However, we found that the SOV score was still being used as a machine learning feature, for example, for protein model quality assessment. Moreover, we tested SOV’99, our SOV_refine, and Q3 on 660 TBM singledomain protein models and found that SOV_refine can indicate the threedimensional quality (indicated by GDTTS and TMscore) of protein models by comparisons at the secondary structure level. Furthermore, we demonstrated SOV score’s usefulness in a newlyemerged bioinformatics problem of inferring TAD locations in mammalian genomes, showing SOV score still could be widely used in bioinformatics research.
Conclusions
In this article, we presented a further modified definition of segment overlap measures (SOV_refine) based on the definition released in 1999 (SOV’99). Specifically, we redefined the assignment of allowance for the overlapping segment pairs. The original definition of allowance has obvious drawbacks and is only based on the overlap level and length of segments. Here we provided a new definition according to the prediction difficulty of reference sequence and the local performance of predicted segments. It can ensure that the amount of allowance is increased when more elements in the segment of a predicted sequence are further predicted accurately.
We performed analysis on the statistical significance of Q3, SOV’99, and SOV_refine and concluded that the probability of finding Q3 ≤ 0.26, SOV’99 ≤ 0.24, and SOV_refine ≤ 0.19 for two random proteins was close to 1. These findings indicate that we can distinguish two protein structures or models at the secondary structure level if their Q3 or SOV scores are greater than or equal to these corresponding thresholds. We can also conclude that for threestate secondary structure when a given score from two random proteins meets Q3 ≤ 0.97 and SOV_refine ≤ 0.94, the two proteins have high probability sharing different CATH folds; for eightstate secondary structure when a given score from two random proteins meets Q8 ≤ 0.95 and SOV_refine ≤ 0.90, the two proteins have high probability sharing different CATH folds. These results also indicate that compared to TM_score we need to get a higher Q3 or SOV_refine scores of any two protein structures or models to determine whether they share the same fold. Meanwhile, we also observed that the twodimensional alignment scores (i.e., Q3, SOV’99, and SOV_refine) are not as effective as the threedimensional alignment scores (i.e., TMscore) when they are used to determine whether two random proteins have the same fold.
We provided another two applications to demonstrate the advantages of SOV_refine compared to SOV’99. One of them is to use SOV scores as features in machinelearning tools for quality assessment of protein models. Our evaluation results show that adding SOV_refine into the basic machine learning feature set results in a larger improvement on performance compared to adding SOV’99. The other application is to use SOV scores as similarity measure for different TAD definitions. The results show that SOV_refine can better distinguish the obvious difference in TAD definitions, whereas SOV’99 often assigns false perfect scores.
We implemented SOV_refine and reimplemented SOV’99 as a standalone computer program. Technically, it can handle unlimited number of states in a reference sequence. However, we highly recommend not to use them when the number of states is quite large (e.g., > 10 states) because more states will reduce the usefulness and significance of SOV scores, in which case the accuracy on a perelement base (e.g., Q3) would be more suitable.
Methods
In this section, we describe the definition of SOV_refine in detail. For the purpose of consistency, we use the same denotations as used in [4]. Here, the reference s_{ r } and predicted s_{ p } sequences are respectively the native and predicted assignments of protein secondary structures in three states (i.e., H, E, and C); however, our program can handle unlimited number of states and optional labels for states, even though it may not make much sense if the number of states is too large.
The SOV score for each state i, SOV(i), is calculated individually, and then the global SOV score is a weighted combination of individual state scores. Let s_{ 1 } be a segment in state i in s_{ r } and s_{ 2 } in state i in s_{ p }. A pair of overlapping segments is denoted as (s_{ 1 }, s_{ 2 }); and the set of these pairs for state i is S(i) = {(s_{1}, s_{2}) s_{1} ⋂ s_{2} ≠ ∅}. If given s_{ 1 }, there are no overlapping segments s_{ 2 }, then we define another set S^{′}(i) = {(s_{1}, ∀ s_{2}) s_{1} ⋂ s_{2} = ∅}. The SOV(i) is defined as follows:
where len(s_{ 1 }) is the number of elements in segment s_{ 1 }; minov(s_{ 1 }, s_{ 2 }) is the number of identical (actually overlapping in istate) elements in both s_{ 1 } and s_{ 2 }, while maxov(s_{ 1 }, s_{ 2 }) is the total number of elements for which either of the two segments is assigned state i; δ(s_{1}, s_{2}) is the amount of allowance assigned to the pair. N(i) is the normalization value defined as:
The new definition of allowance is:
where len(s_{ r }) is the number of elements in s_{ r }; and δ(all) is the total allowance assigned to the whole reference sequence: it can be a fixed value for all reference sequences or depends on each individual sequence. For example, if δ(all) = 1, then all allowance values for segment pairs should be less than or equal to one. Considering that it is difficult to determine a proper fixed value of δ(all), we further define it using the number of states N_{ C } and the length of all segments in s_{ r }:
where N_{ S } is the number of segments in s_{ r }; s_{ j } is the jth segment; λ is an adjustable scale parameter and used to limit the range of δ(all). In the test example for assessing the quality of predicted protein secondary structures, λ equaling to 1 is acceptable. Eq. 4 is designed based on two intuitive facts: (1) More allowance should be assigned when the number of states in s_{ r } is larger because it makes the prediction difficult; (2) More allowance should be assigned when the weighted average length (the denominator part in Eq. 4) of all segments in s_{ r } is smaller because a small average length results in more boundary regions, which increases the difficulty of predictions. In order to avoid more than perfect for SOV(i), when the amount of allowance calculated in Eq. 3 is larger than (maxov(s_{ 1 },s_{ 2 })minov(s_{ 1 },s_{ 2 })) the allowance is set to (maxov(s_{ 1 },s_{ 2 })minov(s_{ 1 },s_{ 2 })).
Suppose that the number of states in s_{ r } is N_{ C }, then the final SOV score, SOV_refine, can be defined as:
The new definition of SOV_refine remedies three deficiencies found in SOV’99. First, the amount of allowance does not have to be an integer. Instead, the amount of allowance defined in Eq. 3 is based on the local performance of s_{ 2 } and a fractional part of δ(all). Second, SOV’99 cannot ensure that the amount of allowance keeps increasing when more residues in a segment in s_{ p } are further predicted accurately, whereas SOV_refine can. Third, we take the allowance for the whole reference sequence s_{ r } into consideration, because sometimes it may be much easier to predict (e.g., when s_{ r } only has one state), while for other cases it may be very difficult (e.g., if s_{ r } has eight states and multiple segments with different lengths). In our design, the value of δ(all) depends on s_{ r }, that is, reference sequences with different lengths and prediction difficulty have different δ(all).
Abbreviations
 3D:

Threedimensional
 Q3:

Three states for protein secondary structure
 QA:

Quality assessment
 SOV:

Segment overlap measures
 TAD:

Topologically Associating Domain
 TBM:

TemplateBased Modelling
References
 1.
Kabsch W, Sander C. Dictionary of protein secondary structure: pattern recognition of hydrogenbonded and geometrical features. Biopolymers. 1983;22(12):2577–637.
 2.
Dixon JR, Selvaraj S, Yue F, Kim A, Li Y, Shen Y, Hu M, Liu JS, Ren B. Topological domains in mammalian genomes identified by analysis of chromatin interactions. Nature. 2012;485(7398):376–80.
 3.
Rost B, Sander C, Schneider R. Redefining the goals of protein secondary structure prediction. J Mol Biol. 1994;235(1):13–26.
 4.
Zemla A, Venclovas Č, Fidelis K, Rost B. A modified definition of Sov, a segmentbased measure for protein secondary structure prediction assessment. Proteins: Structure Function Bioinformatics. 1999;34(2):220–3.
 5.
Wang S, Peng J, Ma J, Xu J. Protein secondary structure prediction using deep convolutional neural fields. Sci Rep. 2016;6:18962.
 6.
Aloy P, Stark A, Hadley C, Russell RB. Predictions without templates: new folds, secondary structure, and contacts in CASP5. Proteins: Structure Function Bioinformatics. 2003;53(S6):436–56.
 7.
Jones D. Protein secondary structure prediction based on positionspecific scoring matrices. J Mol Biol. 1999;292(2):195–202.
 8.
Geourjon C, Deleage G. SOPMA: significant improvements in protein secondary structure prediction by consensus prediction from multiple alignments. Comp Applicat Biosci. 1995;11(6):681–4.
 9.
Kim H, Park H. Protein secondary structure prediction based on an improved support vector machines approach. Protein Eng. 2003;16(8):553–60.
 10.
Ward JJ, McGuffin LJ, Buxton BF, Jones DT. Secondary structure prediction with support vector machines. Bioinformatics. 2003;19(13):1650–5.
 11.
Guermeur Y, Geourjon C, Gallinari P, Del G. Improved performance in protein secondary structure prediction by inhomogeneous score combination. Bioinformatics. 1999;15(5):413–21.
 12.
Pollastri G, Mclysaght A. Porter: a new, accurate server for protein secondary structure prediction. Bioinformatics. 2005;21(8):1719–20.
 13.
Hua S, Sun Z. A novel method of protein secondary structure prediction with high segment overlap measure: support vector machine approach. J Mol Biol. 2001;308(2):397–407.
 14.
Martin J, Letellier G, Marin A, Taly JF, de Brevern AG, Gibrat JF. Protein secondary structure assignment revisited: a detailed analysis of different assignment methods. BMC Struct Biol. 2005;5(1):1.
 15.
Wang Z, Eickholt J, Cheng J. APOLLO: a quality assessment Service for Single and Multiple Protein Models. Bioinformatics. 2011;27(12):1715–6.
 16.
Cao R, Wang Z, Wang Y, Cheng J. SMOQ: a tool for predicting the absolute residuespecific quality of a single protein model with support vector machines. BMC Bioinform. 2014;15(1):120.
 17.
Liu T, Wang Y, Eickholt J, Wang Z. Benchmarking deep networks for predicting residuespecific quality of individual protein models in CASP11. Sci Rep. 2016;6:19301.
 18.
Kryshtafovych A, Barbato A, Monastyrskyy B, Fidelis K, Schwede T, Tramontano A. Methods of model accuracy estimation can help selecting the best models from decoy sets: assessment of model accuracy estimations in CASP11. Proteins: Structure Function Bioinformatics. 2016;84(S1):349–69.
 19.
Zemla A. LGA: a method for finding 3D similarities in protein structures. Nucleic Acids Res. 2003;31(13):3370–4.
 20.
Zhang Y, Skolnick J. TMalign: a protein structure alignment algorithm based on the TMscore. Nucleic Acids Res. 2005;33(7):2302–9.
 21.
Wang S, Ma J, Peng J, Xu J. Protein structure alignment beyond spatial proximity. Sci Rep. 2013;3:1448.
 22.
Frishman D, Argos P. Knowledgebased protein secondary structure assignment. Proteins Struct Funct Genet. 1995;23(4):566–79.
 23.
Xu J, Zhang Y. How significant is a protein structure similarity with TMscore= 0.5? Bioinformatics. 2010;26(7):889–95.
 24.
Chen VB, Arendall WB, Headd JJ, Keedy DA, Immormino RM, Kapral GJ, Murray LW, Richardson JS, Richardson DC. MolProbity: allatom structure validation for macromolecular crystallography. Acta Crystallogr D Biol Crystallogr. 2010;66(1):12–21.
 25.
Sillitoe I, Lewis TE, Cuff A, Das S, Ashford P, Dawson NL, Furnham N, Laskowski RA, Lee D, Lees JG. CATH: comprehensive structural and functional annotations for genome sequences. Nucleic Acids Res. 2014;43(D1):D376–81.
 26.
Magnan CN, Baldi P. SSpro/ACCpro 5: almost perfect prediction of protein secondary structure and relative solvent accessibility using profiles, machine learning and structural similarity. Bioinformatics. 2014;30(18):2592–7.
 27.
Breiman L. Random forests. Mach Learn. 2001;45(1):5–32.
 28.
Dali R, Blanchette M. A critical assessment of topologically associating domain prediction tools. Nucleic Acids Res. 2017;45(6):2994–3005.
Acknowledgements
Not applicable
Funding
Research reported in this publication was supported by a startup funding from the University of Miami to ZW and the National Institute of General Medical Sciences of the National Institutes of Health under award number 1R15GM120650 to ZW.
Availability of data and materials
Not applicable
Author information
Affiliations
Contributions
TL designed, implemented, and benchmarked the algorithm and made the figures and tables. ZW guided and advised the research. TL and ZW wrote the manuscript. Both authors read and approved the final manuscript.
Corresponding author
Correspondence to Zheng Wang.
Ethics declarations
Ethics approval and consent to participate
Not applicable
Consent for publication
Not applicable
Competing interests
The authors declare that they have no competing interests.
Publisher’s Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Open Access This article is distributed under the terms of the Creative Commons Attribution 4.0 International License (http://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made. The Creative Commons Public Domain Dedication waiver (http://creativecommons.org/publicdomain/zero/1.0/) applies to the data made available in this article, unless otherwise stated.
About this article
Received
Accepted
Published
DOI
Keywords
 Segment overlap score
 SOV score
 Protein secondary structure prediction
 Assessment of protein secondary structure predictions
 Protein structure similarity
 Similarity of segmented biological sequences
 Comparing different definitions of topologically associating domains