Nothing Special   »   [go: up one dir, main page]

Next Article in Journal
Dynamic Mixed Data Analysis and Visualization
Next Article in Special Issue
NaRnEA: An Information Theoretic Framework for Gene Set Analysis
Previous Article in Journal
Multicriteria Group Decision Making Based on Intuitionistic Normal Cloud and Cloud Distance Entropy
Previous Article in Special Issue
An Information-Theoretic Bound on p-Values for Detecting Communities Shared between Weighted Labeled Graphs
You seem to have javascript disabled. Please note that many of the page functionalities won't work as expected without javascript enabled.
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Information Theory for Biological Sequence Classification: A Novel Feature Extraction Technique Based on Tsallis Entropy

by
Robson P. Bonidia
1,*,
Anderson P. Avila Santos
1,2,
Breno L. S. de Almeida
1,
Peter F. Stadler
3,
Ulisses Nunes da Rocha
2,*,
Danilo S. Sanches
4 and
André C. P. L. F. de Carvalho
1
1
Institute of Mathematics and Computer Sciences, University of São Paulo, São Carlos 13566-590, Brazil
2
Department of Environmental Microbiology, Helmholtz Centre for Environmental Research-UFZ GmbH, 04318 Leipzig, Germany
3
Department of Computer Science and Interdisciplinary Center of Bioinformatics, University of Leipzig, 04107 Leipzig, Germany
4
Department of Computer Science, Federal University of Technology-Paraná—UTFPR, Cornélio Procópio 86300-000, Brazil
*
Authors to whom correspondence should be addressed.
Entropy 2022, 24(10), 1398; https://doi.org/10.3390/e24101398
Submission received: 11 August 2022 / Revised: 16 September 2022 / Accepted: 24 September 2022 / Published: 1 October 2022
(This article belongs to the Special Issue Information Theory in Computational Biology)

Abstract

:
In recent years, there has been an exponential growth in sequencing projects due to accelerated technological advances, leading to a significant increase in the amount of data and resulting in new challenges for biological sequence analysis. Consequently, the use of techniques capable of analyzing large amounts of data has been explored, such as machine learning (ML) algorithms. ML algorithms are being used to analyze and classify biological sequences, despite the intrinsic difficulty in extracting and finding representative biological sequence methods suitable for them. Thereby, extracting numerical features to represent sequences makes it statistically feasible to use universal concepts from Information Theory, such as Tsallis and Shannon entropy. In this study, we propose a novel Tsallis entropy-based feature extractor to provide useful information to classify biological sequences. To assess its relevance, we prepared five case studies: (1) an analysis of the entropic index q; (2) performance testing of the best entropic indices on new datasets; (3) a comparison made with Shannon entropy and (4) generalized entropies; (5) an investigation of the Tsallis entropy in the context of dimensionality reduction. As a result, our proposal proved to be effective, being superior to Shannon entropy and robust in terms of generalization, and also potentially representative for collecting information in fewer dimensions compared with methods such as Singular Value Decomposition and Uniform Manifold Approximation and Projection.

1. Introduction

The accelerated evolution of sequencing technologies has generated significant growth in the number of sequence data [1], opening up new opportunities and creating new challenges for biological sequence analysis. To take advantage of the increased predictive power of machine learning (ML) algorithms, recent works have investigated the use of these algorithms to analyze biological data [2,3].
The development of effective methods for sequence analysis, through ML, benefits the research advancement in new applications [4,5], such as understanding several problems [4,5], e.g., cancer diagnostics [6], development of CRISPR-Cas systems [7], drug discovery and development [8] and COVID-19 diagnosis [9]. Nevertheless, ML algorithms applied to the analysis of biological sequences present challenges, such as feature extraction [10]. For non-structured data, as is the case of biological sequences, feature extraction is a key step for the success of ML applications [11,12,13].
Previous works have shown that universal concepts from Information Theory (IT), originally proposed by Claude Shannon (1948) [14], can be used to extract relevant information from biological sequences [15,16,17]. According to [18], an IT-based analysis of symbolic sequences is of interest in various study areas, such as linguistics, biological sequence analysis or image processing, whose relevant information can be extracted, for example, by Shannon’s uncertainty theory [19].
Studies have investigated the analysis of biological sequences with Shannon entropy in a wide range of applications [19,20,21]. Given their large applicability, according to [22], it is important to explore the possibility of generalized entropies, such as Tsallis [23,24], which was proposed to generalize the Boltzmann/Gibbs’s traditional entropy to non-extensive physical systems [25]. This class of generalized entropy has been used for different problems, e.g., image analysis [25,26], inference of gene regulatory networks [27], DNA analysis [20] induction of decision trees [28] and classification of epileptic seizures [29].
In [25], the authors proposed a new image segmentation method using Tsallis entropy. Later, Ref. [26] showed a novel numerical approach to calculate the Tsallis entropic index feature for a given image. In [27], the authors introduced the use of generalized entropy for the inference of gene regulatory networks. DNA analysis using entropy (Shannon, Rényi and Tsallis) and phase plane concepts was presented in [20], while [28] used the concept of generalized entropy for decision trees. Recently, Ref. [29] investigated a novel single feature based on Tsallis entropy to classify epileptic seizures. These studies report a wide range of contributions to the use of Tsallis entropy in different domains. To the best of our knowledge, this paper is the first work proposing its use as a feature (feature extraction) to represent distinct biological sequences. Additionally, it presents the first study of different Tsallis entropic indexes and their effects on classical classifiers.
A preliminary version of this proposal was presented in [5]. Due to the favorable results obtained, we created a code to extract different descriptors available in a new programming package, called MathFeature [13], which implements mathematical descriptors for biological sequences. However, until now, we have not studied Tsallis entropy in depth, e.g., its effect, its application to other biological sequence datasets and its comparison with other entropy-based descriptors, e.g., Shannon. Thus, in this paper, we investigate the answers to the following questions:
  • Question 1 (Q1): Are Tsallis entropy-based features robust for extracting information from biological sequences in classification problems?
  • Question 2 (Q2): Does the entropic index affect the classification performance?
  • Question 3 (Q3): Is Tsallis entropy as robust as Shannon entropy for extracting information from biological sequences?
We are evaluating robustness in terms of performance, e.g., accuracy, recall and F1 score, of the feature vectors extracted by our proposal on different biological sequence datasets. Finally, this study makes the following main research contributions: we propose an effective feature extraction technique based on Tsallis entropy, being robust in terms of generalization, and also potentially representative for collecting information in fewer dimensions for sequence classification problems.

2. Literature Review

In this section, we develop a systematic literature review to present and summarize feature extraction descriptors for biological sequences (DNA, RNA, or protein). This review aims to report the need and lack of studies with mathematical descriptors, such as entropy, evidencing the contribution of this article. This section followed the Systematic Literature Review (SLR) Guidelines in Software Engineering [30], which, according to [30,31], allows a rigorous and reliable evaluation of primary studies within a specific topic. We base our review on recommendations from previous studies [30,31,32].
We propose to address the following problem: How can we numerically represent a biological sequence (such as DNA, RNA, or protein) in a numeric vector that can effectively reflect the most discriminating information in a sequence? To answer this question, we reviewed ML-based feature extraction tools (or packages, web servers and toolkits) that aim, as a proposal, to provide several feature descriptors for biological sequences—that is, without a defined scope, and, therefore, generalist studies. Moreover, we used the following electronic databases: ACM Digital Library, IEEE Xplore Digital Library, PubMed and Scopus. We chose the Boolean method [33] to search primary studies in the literature databases. The standard search string was: (“feature extraction” OR “extraction” OR “features” OR “feature generation” OR “feature vectors”) AND (“machine” OR “learning”) AND (“tool” OR “web server” OR “package” OR “toolkit”) AND (“biological sequence” OR “sequence”).
Due to different query languages and limitations between the scientific article databases, there were some differences in the search strings. Therefore, our first step was to apply search keys to all databases, returning a set of 1404 studies. Furthermore, we used the Parsifal tool to assist our review and obtain better accuracy and reliability. Thereafter, duplicate studies were removed, returning an amount of 1097 titles (307 duplicate studies).
Then, we performed a thorough analysis of the titles, keywords and abstracts, according to inclusion and exclusion criteria: (1) Studies in English, (2) Studies with different feature extraction techniques, (3) Studies with generalist tools and (4) Studies published in journals. We accepted 28 studies (we rejected, 1069). Finally, after pre-selecting the studies, we performed a data synthesis, to apply an assessment based on the quality criteria: (1) Are the study aims specified? (2) Study with different proposals/results? (3) Study with complete results?
Hence, of the 28 studies, 3 were eliminated, leading to a final set of 25 studies (see Supplementary Table S1). As previously mentioned, we assessed generalist tools for feature extraction, since this type of study would provide several descriptors, presenting an overview of ways to numerically represent biological sequences (which would not be possible by evaluating studies dedicated to some specific problem). As expected, we found more than 100 feature descriptors. We chose to divide them into large groups (16 groups—these were defined based on all studies), as shown in Supplementary Table S2. Then, we created Table 1 with all the feature descriptors found in the 25 studies (see the complete table in Supplementary Table S3). As can be seen, no study provides mathematical descriptors, such as Tsallis entropy, reinforcing the contribution of our proposal.

3. Information Theory and Entropy

According to [34], IT can be defined as a mathematical treatment of the concepts, parameters, and rules related to the transmission and processing of information. The IT concept was first proposed by Claude Shannon (1948) in the work entitled “A Mathematical Theory of Communication” [14], where he showed how information could be quantified with absolute precision. The entropy originating from IT can be considered a measure of order and disorder in a dynamic system [14,25]. However, to define information and entropy, it is necessary to understand random variables, which, in probability theory, is a mathematical object that can take on a finite number of different states x 1 , , x n with previously defined probabilities p 1 , , p n [35]. According to [5], for a discrete random variable R taking values in { r [ 0 ] , r [ 1 ] , r [ 2 ] , , r [ N 1 ] } with probabilities { p [ 0 ] , p [ 1 ] , p [ 2 ] , , p [ N 1 ] } , represented as P ( R = r [ n ] ) = p [ n ] , we can define self-information or information as [26]
I = l o g ( p ) .
Thus, the Shannon entropy H S is defined by
H S = n = 0 N 1 p [ n ] l o g 2 p [ n ] .
Here, N is the number of possible events and p [ n ] the probability that event n occurs. Fundamentally, with Shannon entropy, we can reach a single value that quantifies the information contained in different observation periods [36]. Furthermore, it is important to highlight that the Boltzmann/Gibbs entropy was redefined by Shannon as a measure of uncertainty [25]. This formalism, known as Boltzmann–Gibbs–Shannon (BGS) statistics, has often been used to interpret discrete and symbolic data [18]. Moreover, according to [25,37], if we decompose a physical system into two independent statistical subsystems A and B, the Shannon entropy has the extensive property (additivity)
H S ( A + B ) = H S ( A ) + H S ( B )
According to [38], complementary information on the importance of specific events can be generated using the notion of generalized entropy, e.g., outliers or rare events. Along these lines, Constantino Tsallis [23,24] proposed a generalized entropy of the BGS statistics, which can be defined as follows:
H T = 1 q 1 1 n = 0 N 1 p [ n ] q .
Here, q is called the entropic index, which, depending on its value, can represent various types of entropy. Depending on the value of q, three different entropies can be defined [25,37]:
  • Superextensive entropy ( q < 1 ) :
    H T ( A + B ) < H T ( A ) + H T ( B )
  • Extensive entropy ( q = 1 ) :
    H T ( A + B ) = H T ( A ) + H T ( B )
  • Subextensive entropy ( q > 1 ) :
    H T ( A + B ) > H T ( A ) + H T ( B )
When q < 1 , the Tsallis entropy is superextensive; for q = 1 , it is extensive (e.g., leads to the Shannon entropy), and for q > 1 , it is subextensive [39]. Therefore, based on these differences, it is important to explore the possibility of generalized entropies [22,28,40]. Another notable generalized entropy is the Rényi entropy, which generalizes the Shannon entropy, the Hartley entropy, the collision entropy and the min-entropy [41,42]. The Rényi entropy can be defined as follows:
H R = 1 1 q l o g 2 n = 0 N 1 p [ n ] q .
As in the Tsallis entropy, q = 1 leads to Shannon entropy.

4. Materials and Methods

In this section, we describe the experimental methodology adopted for this study, which is divided into five stages: (1) data selection; (2) feature extraction; (3) extensive analysis of the entropic index; (4) performance analysis; (5) comparative study.

4.1. A Novel Feature Extraction Technique

Our proposal is based on the studies of [5,20]. To generate our probabilistic experiment [15], we use a known tool in biology, the k-mer. In this method, each sequence is mapped in the frequency of neighboring bases k, generating statistical information. The k-mer is denoted in this work by P k , corresponding to Equation (9).
P k ( s ) = c i k N k + 1 = ( c 1 1 N 1 + 1 , , c 4 1 N 1 + 1 , c 4 + 1 2 N 2 + 1 , , c i k N k + 1 ) k = 1 , 2 , , n .
Here, each sequence ( s ) was assessed with frequencies of k = 1 , 2 , , 24 , in which c i k is the number of occurrences with length k in a sequence ( s ) with length N; the index i { 1 , 2 , , 4 1 + + 4 k } refers to an analyzed substring (e.g., [ { A A A A } , , { T T T T } ] , for k = 4 ). Here, after counting the absolute frequencies of each k, we generate relative frequencies and then apply Tsallis entropy to generate the features. In the case of protein sequences, index i is { 1 , 2 , , 20 1 + + 20 k } . For a better understanding, Algorithm 1 demonstrates our pseudocode.
Algorithm 1: Pseudocode of the Proposed Technique
Entropy 24 01398 i001
This algorithm is divided into five steps: (1) each sequence is mapped to k m e r s ; (2) extraction of the absolute frequency of each k m e r ; (3) extraction of the relative frequency of each k m e r based on absolute frequency; (4) extraction of the Tsallis entropy, based on the relative frequency for each k m e r —see Equation (4); (5) generation, for each k m e r , of an entropic measure. Regarding interpretability, each entropic measure represents a k m e r , e.g., 1-mer = frequency of A, C, T, G. In other words, analyzing the best measures—for example, through a feature importance analysis—we can determine which k m e r s are more relevant to the problem under study, providing an indication of which combination of nucleotides or amino acids contributes to the classification of the sequences.

4.2. Benchmark Dataset and Experimental Setting

To validate the proposal, we divided our experiments into five case studies:
  • Case Study I: Assessment of the Tsallis entropy and the effect of the entropic index q, generating 100 feature vectors for each benchmark dataset with 100 different q parameters (entropic index). The features were extracted by Algorithm 1, with q varying from 0.1 to 10.0 in steps of 0.1 (except 1.0, which leads to the Shannon entropy). The goal was to find the best values for the parameter q to be used in the experiments. For this, three benchmark datasets from previous studies were used [5,43,44]. For the first dataset (D1), the selected task was long non-coding RNAs (lncRNA) vs. protein-coding genes (mRNA), as in [45], using a set with mRNA and lncRNA sequences (500 for each label—benchmark dataset [5]). For the second dataset (D2), a benchmark set from [5], the selected task was the induction of a classifier to distinguish circular RNAs (cirRNAs) from other lncRNAs using 1000 sequences (500 for each label). The third dataset (D3) is for Phage Virion Protein (PVP) classification, from [44], with 129 PVP and 272 non-PVP sequences.
  • Case Study II: We use the best parameters ( q : entropic index—found in case study I) to evaluate its performance on new datasets: D4—Sigma70 Promoters [46] (2141 sequences), D5—Anticancer Peptides [47] (344 sequences) and D6—Severe Acute Respiratory Syndrome Coronavirus 2 (SARS-CoV-2, 24815 sequences) [13].
  • Case Study III—Comparing Tsallis with Shannon Entropy: As a baseline of the comparison between methods, we use Shannon entropy, as we did not find any article studying the form of proposed classification with Tsallis entropy and the effect of the entropic parameter with different classifiers. In this experiment, we use D1, D2, D3, D4, D5 and D6.
  • Case Study IV—Comparing Generalized Entropies: To better understand the effectiveness of generalized entropies for feature extraction, we evaluated Tsallis with the Rényi entropy. In this case, the evaluations of the two approaches were conducted by using the experiments from case study I, changing the entropic index for generating the datasets from 0.1 to 10.0 in steps of 0.1, and inducing the CatBoost classifier. In addition, the datasets used were D1, D2 and D3.
  • Case Study V—Dimensionality Reduction Analysis: Finally, we assessed our proposal with other known techniques of feature extraction and dimensionality reduction, e.g., Singular Value Decomposition (SVD) [48] and Uniform Manifold Approximation and Projection (UMAP) [49], using datasets D1, D2, D3 and D5. We also added three new benchmark datasets provided by [50] to predict recombination spots (D7) with 1050 sequences (it contained 478 positive sequences and 572 negative sequences) and for the HIV-1 M pure subtype against CRF classification (D8) with 200 sequences (it contained 100 positive and negative sequences) [51]. In addition, we also used a multiclass dataset (D9) containing seven bacterial phyla with 488 small RNA (sRNA), 595 transfer RNA (tRNA) and 247 ribosomal RNA (rRNA) from [52]. Moreover, to apply SVD and UMAP, we kept the same feature descriptor by k-mer frequency.
For data normalization in all stages, we used the min–max algorithm. Furthermore, we investigated five classification algorithms, such as Gaussian Naive Bayes (GaussianNB), Random Forest (RF), Bagging, Multi-Layer Perceptron (MLP) and CatBoost. To induce our models, we randomly divided the datasets into ten separate sets to perform 10-fold cross-validation (case study I and case study V) and hold-out (70% of samples for training and 30% for testing—case study II, case study III, and case study IV). Finally, we assessed the results with accuracy (ACC), balanced accuracy (BACC), recall, F1 score and Area Under the Curve (AUC). In D9, we considered metrics suitable for multiclass evaluation.

5. Results and Discussion

5.1. Case Study I

As aforementioned, we induced our classifiers (using 10-fold cross-validation) across all feature vectors generated with 100 different q parameters (totaling 300 vectors (3 datasets times 100 parameters)). Thereby, we obtained the results presented in Table 2. This table shows the best and worst parameter (entropic parameter q) of each algorithm in the three benchmark datasets, taking into account the ACC metric.
Thereby, evaluating each classifier, we observed that the CatBoost performed best in all datasets, with 0.9440 ( q = 2.3 ), 0.8300 ( q = 4.0 ), 0.7282 ( q = 1.1 ) in D1, D2 and D3, respectively. The other best classifiers were RF, with 0.9430 (q = 0.4 − D1) and 0.8220 (q = 5.3 − D2), followed by Bagging, MLP, and GaussianNB. Furthermore, in general, we noticed that the best results presented parameters between 1.1 < q < 5.0 , i.e., when the Tsallis entropy was subextensive. Along the same lines, it can be observed in Table 2 that the worst parameters are between 9.0 < q < 10.0 , when the Tsallis entropy is also subextensive. However, for a more reliable analysis, we plotted graphs with the results of all tested parameters ( 0.1 to 10.0 in steps of 0.1 ), as shown in Figure 1.
A large difference can be observed in the entropy obtained by each parameter q, mainly in benchmark D3. Thereby, analyzing D1 and D2, we noticed a pattern of robust results until q = 6 , for the best classifiers in both datasets. However, as the q parameter increases, the classifiers are less accurate. On the other hand, if we look at D3, the entropy obtained for each parameter q presents a much greater variation, but following the same drop with parameters close to q = 10 . Regarding the superextensive entropy ( q < 1 ) , some cases showed robust results; however, most classifiers behaved better with the subextensive entropy.

5.2. Case Study II

After substantially evaluating the entropic index, our findings indicated that the best parameters were among 1.1 < q < 5.0 . Thereby, we generated new experiments using five parameters to test their efficiency in new datasets, with q = ( 0.5 , 2.0 , 3.0 , 4.0 , 5.0 ) , as shown in Table 3 (sigma70 promoters—D4), Table 4 (anticancer peptides—D5) and Table 5 (SARS-CoV-2—D6). Here, we generated the results with the two best classifiers (RF and Catboost—best in bold).
Assessing each benchmark dataset, we note that the best results were of ACC: 0.6687 and AUC: 0.6108 in D4 (RF, q = 2.0 ), ACC: 0.7212 and AUC: 0.7748 in D5 (RF, q = 3.0 ), and ACC: 1.0000 and AUC: 1.0000 in D5 (RF and CatBoost, q = 5.0 ). Once more, the results confirm that the best parameters are in the range of 1.1 < q < 5.0 , indicating a good choice when using Tsallis entropy. The perfect classification at D6 is supported by other studies in the literature [53,54,55]. Nevertheless, after testing the Tsallis entropy on six benchmark datasets, we noticed an indication that this approach behaves better with longer sequences, e.g., D1 (mean length ≈ 751 bp), D2 (mean length ≈ 2799 bp), and D6 (mean length ≈ 10,870 bp) showed robust results, while D3 (mean length ≈ 268 bp), D4 (mean length ≈ 81 bp), and D5 (mean length ≈ 26 bp) showed less accurate results. Nonetheless, Tsallis entropy could contribute to hybrid approaches, as our proposal achieved relevant results in four datasets.

5.3. Case Study III—Comparing Tsallis with Shannon Entropy

Here, we used Shannon entropy as a baseline for comparison, according to Table 6. Various studies have covered the biological sequence analysis with Shannon entropy, in the most diverse applications. For a fair analysis, we reran the experiments on all datasets (case study I and II, six datasets), using hold-out, with the same train and test partition for both approaches. Once more, we used the best classifiers in case study II (RF and CatBoost), but, for a better understanding, we only show the best result in each dataset.
According to Table 6, our proposal with Tsallis entropy showed better results of ACC (5 wins), recall (4 wins), F1 score (5 wins), and BACC (5 wins) than Shannon entropy in five datasets, falling short only on D6, with a small difference of 0.0002 . Analyzing each metric individually, we observed that the best Tsallis parameters resulted in an F1 score gain compared to Shannon entropy of 5.29 % and 1.81 % in D4 and D5, respectively. Other gains were repeated in ACC, recall, and BACC. In the overall average, our proposal achieved improvements of 0.51%, 1.52%, 1.34%, and 0.62% in ACC, recall, F1 score, and BACC, respectively. Despite a lower accuracy in D3 and D4, this approach alone delivered a BACC of 0.6342 and 0.5845, i.e., it is a supplementary methodology to combine with other feature extraction techniques available in the literature. Based on this, we can state that Tsallis entropy is as robust as Shannon entropy for extracting information from biological sequences.

5.4. Case Study IV—Comparing Generalized Entropies

According to the Tsallis entropy results, wherein it overcame Shannon entropy, we realized the strong performance of generalized entropy as a feature descriptor for biological sequences. For this reason, we also evaluated the influence of another form of generalized entropy, such as Rényi entropy [42], as a good feature descriptor for biological sequences. Here, we investigated the performance of Tsallis and Rényi entropy, changing the entropic index for D1, D2, and D3. Moreover, we have chosen the best classifier from case study I (CatBoost).
When considering the same reproducible environment for the experiment, the performance peak was the same for both methods, as we can see in Figure 2, with graphs containing accuracy performance results for all the entropic index values (from 0.1 to 10.0). Regarding the best classification performance, for D1 (Figure 2a), we had ACC: 0.9600, recall: 0.9667, F1 score: 0.9603, and BACC: 0.9600; for D2 (Figure 2b), we obtained ACC: 0.8300, recall: 0.7733, F1 score: 0.8198, and BACC: 0.8300; and for D3 (Figure 2c), we had ACC: 0.7521, recall: 0.359, F1 score: 0.4828, and BACC: 0.649. As seen earlier, Tsallis entropy performs poorly from a specific entropy index onwards, but Rényi entropy demonstrates more consistent performance when compared to Tsallis, representing a possible alternative. Nevertheless, the results again highlight the promising use of generalized entropies as a feature extraction approach for biological sequences.

5.5. Case Study V—Dimensionality Reduction

In this last case study, we compared our proposal with other known techniques for feature extraction and dimensionality reduction in the literature, using the same representation of the biological sequences, the k m e r frequency. In particular, for each DNA/RNA sequence, we generated k m e r s from k = 1 to k = 10 , while, for proteins, we generated it until k = 5 , considering the high number of combinations with amino acids. All datasets used have around 1000 biological sequences, considering the prohibitive computational cost to deal with the k m e r approach. In this study, our objective was to use SVD and UMAP to reduce the dimensionality of the k m e r feature vector by extracting new features, as we did in our approach. However, high values of k present high computational costs, due to the amount of generated features, e.g., k = 6 in DNA (4096 features) and k = 3 in protein (8000 features).
From previous case studies, we realized that the feature extraction with Tsallis entropy provided interesting results. Thereby, we extended our study, applying SVD and UMAP in the datasets with k m e r frequencies, reducing them to 24 components, comparable to the dimensions generated in our studies. Fundamentally, UMAP can deal with sparse data, as can SVD, which is known for its efficiency in dealing with this type of data [56,57,58]. Both reduction methods can be used in the context of working with high-dimensional data. Although UMAP is widely used for visualization [59,60], the reduction method can be used for feature extraction, which is part of an ML pipeline [61]. UMAP can also be used with raw data, without needing to adopt another reduction technique before using it [58]. We induced the CatBoost classifier using 10-fold cross-validation. We obtained the results listed in Table 7.
As can be seen, Tsallis entropy achieved five wins, against two for SVD and zero for UMAP, taking into account the ACC. In addition, in the general average, we obtained a gain of more than 18% in relation to SVD and UMAP in ACC, indicating that our approach can be potentially representative for collecting information in fewer dimensions for sequence classification problems.

6. Conclusions

In this study, we evaluated the Tsallis entropy as a feature extraction technique, where we considered five case studies with nine benchmark datasets of sequence classification problems, as follows: (1) we assessed the Tsallis entropy and the effect of the entropic index; (2) we used the best parameters on new datasets; (3–4) we validated our study, using the Shannon and Rényi entropy as a baseline; and (5) we compared Tsallis entropy with other feature extraction techniques based on dimensionality reduction. In all case studies, we found that our proposal is robust for extracting information from biological sequences. Furthermore, the Tsallis entropy’s performance is strongly associated with the length of sequences, providing better results when applied in longer sequences. The experiments also showed that Tsallis entropy is robust when compared to Shannon entropy. Regarding the limitations, we found that the entropic index (q) affects the performance of ML models, particularly when poorly parameterized. Finally, we highlighted good performance for the entropic index with q values between 1.1 and 5.0 .

Supplementary Materials

The following supporting information can be downloaded at: https://www.mdpi.com/article/10.3390/e24101398/s1, Table S1: Overview of Selected Studies. Table S2: Descriptor Group. Table S3: Feature Descriptors Found in All Studies.

Author Contributions

Conceptualization, R.P.B., B.L.S.d.A. and A.P.A.S.; methodology, R.P.B., B.L.S.d.A. and A.P.A.S.; validation, R.P.B., B.L.S.d.A., A.P.A.S., P.F.S., U.N.d.R., D.S.S. and A.C.P.L.F.d.C.; formal analysis, R.P.B. and A.C.P.L.F.d.C.; data curation, R.P.B., P.F.S., U.N.d.R.; writing—original draft preparation, R.P.B. and A.P.A.S.; writing—review and editing, R.P.B., B.L.S.d.A., A.P.A.S., P.F.S., U.N.d.R., D.S.S. and A.C.P.L.F.d.C.; supervision, U.N.d.R., D.S.S. and A.C.P.L.F.d.C. All authors have read and agreed to the published version of the manuscript.

Funding

This project was supported by the Coordenação de Aperfeiçoamento de Pessoal de Nível Superior (CAPES)—Finance Code 001, Google (LARA—2021)—and the Universidade de São Paulo (USP) and São Paulo Research Foundation (FAPESP)—grants #2013/07375-0, #2021/08561-8.

Data Availability Statement

Feature Extraction Technique based on Tsallis Entropy: https://github.com/Bonidia/TsallisEntropyhttps://bonidia.github.io/MathFeature/entropy.html accessed on 10 August 2022.

Acknowledgments

The authors would like to thank USP, CAPES, Google (LARA—2021), CNPq, and FAPESP for their financial support of this research.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Hashemi, F.S.G.; Ismail, M.R.; Yusop, M.R.; Hashemi, M.S.G.; Shahraki, M.H.N.; Rastegari, H.; Miah, G.; Aslani, F. Intelligent mining of large-scale bio-data: Bioinformatics applications. Biotechnol. Biotechnol. Equip. 2018, 32, 10–29. [Google Scholar] [CrossRef] [Green Version]
  2. Silva, J.C.F.; Teixeira, R.M.; Silva, F.F.; Brommonschenkel, S.H.; Fontes, E.P. Machine learning approaches and their current application in plant molecular biology: A systematic review. Plant Sci. 2019, 284, 37–47. [Google Scholar] [CrossRef]
  3. Greener, J.G.; Kandathil, S.M.; Moffat, L.; Jones, D.T. A guide to machine learning for biologists. Nat. Rev. Mol. Cell Biol. 2022, 23, 40–55. [Google Scholar] [CrossRef]
  4. Lou, H.; Schwartz, M.; Bruck, J.; Farnoud, F. Evolution of k-mer frequencies and entropy in duplication and substitution mutation systems. IEEE Trans. Inform. Theor. 2019, arXiv:1812.02250. [Google Scholar] [CrossRef] [Green Version]
  5. Bonidia, R.P.; Sampaio, L.D.H.; Domingues, D.S.; Paschoal, A.R.; Lopes, F.M.; de Carvalho, A.C.P.L.F.; Sanches, D.S. Feature extraction approaches for biological sequences: A comparative study of mathematical features. Brief. Bioinform. 2021, 22, bbab011. [Google Scholar] [CrossRef]
  6. Maros, M.E.; Capper, D.; Jones, D.T.; Hovestadt, V.; von Deimling, A.; Pfister, S.M.; Benner, A.; Zucknick, M.; Sill, M. Machine learning workflows to estimate class probabilities for precision cancer diagnostics on DNA methylation microarray data. Nat. Protoc. 2020, 15, 479–512. [Google Scholar] [CrossRef]
  7. Eitzinger, S.; Asif, A.; Watters, K.E.; Iavarone, A.T.; Knott, G.J.; Doudna, J.A.; Minhas, F. Machine learning predicts new anti-CRISPR proteins. Nucl. Acids Res. 2020, 48, 4698–4708. [Google Scholar] [CrossRef]
  8. Vamathevan, J.; Clark, D.; Czodrowski, P.; Dunham, I.; Ferran, E.; Lee, G.; Li, B.; Madabhushi, A.; Shah, P.; Spitzer, M.; et al. Applications of machine learning in drug discovery and development. Nat. Rev. Drug Discov. 2019, 18, 463–477. [Google Scholar] [CrossRef]
  9. Abubaker Bagabir, S.; Ibrahim, N.K.; Abubaker Bagabir, H.; Hashem Ateeq, R. Covid-19 and Artificial Intelligence: Genome sequencing, drug development and vaccine discovery. J. Infect. Public Health 2022, 15, 289–296. [Google Scholar] [CrossRef]
  10. Storcheus, D.; Rostamizadeh, A.; Kumar, S. A survey of modern questions and challenges in feature extraction. In Proceedings of the Feature Extraction: Modern Questions and Challenges, Montreal, QC, Canada, 11 December 2015; pp. 1–18. [Google Scholar]
  11. Iuchi, H.; Matsutani, T.; Yamada, K.; Iwano, N.; Sumi, S.; Hosoda, S.; Zhao, S.; Fukunaga, T.; Hamada, M. Representation learning applications in biological sequence analysis. Comput. Struct. Biotechnol. J. 2021, 19, 3198–3208. [Google Scholar] [CrossRef]
  12. Cui, F.; Zhang, Z.; Zou, Q. Sequence representation approaches for sequence-based protein prediction tasks that use deep learning. Brief. Funct. Genom. 2021, 20, 61–73. [Google Scholar] [CrossRef]
  13. Bonidia, R.P.; Domingues, D.S.; Sanches, D.S.; de Carvalho, A.C. MathFeature: Feature extraction package for DNA, RNA and protein sequences based on mathematical descriptors. Brief. Bioinform. 2022, 23, bbab434. [Google Scholar] [CrossRef]
  14. Shannon, C.E. A mathematical theory of communication. Bell Syst. Tech. J. 1948, 27, 379–423. [Google Scholar] [CrossRef] [Green Version]
  15. Vinga, S. Information theory applications for biological sequence analysis. Brief. Bioinform. 2013, 15, 376–389. [Google Scholar] [CrossRef] [Green Version]
  16. Pritišanac, I.; Vernon, R.M.; Moses, A.M.; Forman Kay, J.D. Entropy and information within intrinsically disordered protein regions. Entropy 2019, 21, 662. [Google Scholar] [CrossRef] [Green Version]
  17. Vopson, M.M.; Robson, S.C. A new method to study genome mutations using the information entropy. Phys. A Statist. Mech. Appl. 2021, 584, 126383. [Google Scholar] [CrossRef]
  18. Ré, M.A.; Azad, R.K. Generalization of entropy based divergence measures for symbolic sequence analysis. PLoS ONE 2014, 9, e0093532. [Google Scholar] [CrossRef] [Green Version]
  19. Akhter, S.; Bailey, B.A.; Salamon, P.; Aziz, R.K.; Edwards, R.A. Applying Shannon’s information theory to bacterial and phage genomes and metagenomes. Sci. Rep. 2013, 3, 1033. [Google Scholar] [CrossRef] [Green Version]
  20. Machado, J.T.; Costa, A.C.; Quelhas, M.D. Shannon, Rényie and Tsallis entropy analysis of DNA using phase plane. Nonlinear Anal. Real World Appl. 2011, 12, 3135–3144. [Google Scholar] [CrossRef]
  21. Tripathi, R.; Patel, S.; Kumari, V.; Chakraborty, P.; Varadwaj, P.K. Deeplnc, a long non-coding rna prediction tool using deep neural network. Netw. Model. Anal. Health Inform. Bioinform. 2016, 5, 21. [Google Scholar] [CrossRef]
  22. Yamano, T. Information theory based on nonadditive information content. Phys. Rev. E 2001, 63, 046105. [Google Scholar] [CrossRef] [Green Version]
  23. Tsallis, C. Possible generalization of Boltzmann-Gibbs statistics. J. Stat. Phys. 1988, 52, 479–487. [Google Scholar] [CrossRef]
  24. Tsallis, C.; Mendes, R.; Plastino, A.R. The role of constraints within generalized nonextensive statistics. Phys. A Stat. Mech. Appl. 1998, 261, 534–554. [Google Scholar] [CrossRef]
  25. De Albuquerque, M.P.; Esquef, I.A.; Mello, A.G. Image thresholding using Tsallis entropy. Pattern Recognit. Lett. 2004, 25, 1059–1065. [Google Scholar] [CrossRef]
  26. Ramírez-Reyes, A.; Hernández-Montoya, A.R.; Herrera-Corral, G.; Domínguez-Jiménez, I. Determining the entropic index q of Tsallis entropy in images through redundancy. Entropy 2016, 18, 299. [Google Scholar] [CrossRef] [Green Version]
  27. Lopes, F.M.; de Oliveira, E.A.; Cesar, R.M. Inference of gene regulatory networks from time series by Tsallis entropy. BMC Syst. Biol. 2011, 5, 61. [Google Scholar] [CrossRef] [Green Version]
  28. De la Cruz-García, J.S.; Bory-Reyes, J.; Ramirez-Arellano, A. A Two-Parameter Fractional Tsallis Decision Tree. Entropy 2022, 24, 572. [Google Scholar] [CrossRef]
  29. Thilagaraj, M.; Rajasekaran, M.P.; Kumar, N.A. Tsallis entropy: As a new single feature with the least computation time for classification of epileptic seizures. Clust. Comput. 2019, 22, 15213–15221. [Google Scholar] [CrossRef] [Green Version]
  30. Keele, S. Guidelines for Performing Systematic Literature Reviews in Software Engineering; Version 2.3 EBSE Technical Report; EBSE-2007-01; University of Durham: Durham, UK, 2007. [Google Scholar]
  31. Brereton, P.; Kitchenham, B.A.; Budgen, D.; Turner, M.; Khalil, M. Lessons from applying the systematic literature review process within the software engineering domain. J. Syst. Softw. 2007, 80, 571–583. [Google Scholar] [CrossRef] [Green Version]
  32. Kitchenham, B.; Brereton, O.P.; Budgen, D.; Turner, M.; Bailey, J.; Linkman, S. Systematic literature reviews in software engineering—A systematic literature review. Inform. Softw. Technol. 2009, 51, 7–15. [Google Scholar] [CrossRef]
  33. Karimi, S.; Pohl, S.; Scholer, F.; Cavedon, L.; Zobel, J. Boolean versus ranked querying for biomedical systematic reviews. BMC Med. Inform. Decis. Mak. 2010, 10, 58. [Google Scholar] [CrossRef] [Green Version]
  34. Martignon, L. Information Theory. In International Encyclopedia of the Social & Behavioral Sciences; Smelser, N.J., Baltes, P.B., Eds.; Pergamon: Oxford, UK, 2001; pp. 7476–7480. [Google Scholar] [CrossRef]
  35. Adami, C. The use of information theory in evolutionary biology. Ann. N. Y. Acad. Sci. 2012, 1256, 49–65. [Google Scholar] [CrossRef] [Green Version]
  36. Lesne, A. Shannon entropy: A rigorous notion at the crossroads between probability, information theory, dynamical systems and statistical physics. Math. Struct. Comput. Sci. 2014, 24, e240311. [Google Scholar] [CrossRef] [Green Version]
  37. Zhang, Y.; Wu, L. Optimal multi-level thresholding based on maximum Tsallis entropy via an artificial bee colony approach. Entropy 2011, 13, 841–859. [Google Scholar] [CrossRef] [Green Version]
  38. Maszczyk, T.; Duch, W. Comparison of Shannon, Renyi and Tsallis entropy used in decision trees. In Proceedings of the International Conference on Artificial Intelligence and Soft Computing, Zakopane, Poland, 22–26 June 2008; pp. 643–651. [Google Scholar]
  39. Tsallis, C. Nonextensive statistics: Theoretical, experimental and computational evidences and connections. Braz. J. Phys. 1999, 29, 1–35. [Google Scholar] [CrossRef]
  40. Dérian, N.; Pham, H.P.; Nehar-Belaid, D.; Tchitchek, N.; Klatzmann, D.; Eric, V.; Six, A. The Tsallis generalized entropy enhances the interpretation of transcriptomics datasets. PLoS ONE 2022, 17, e0266618. [Google Scholar] [CrossRef]
  41. Fehr, S.; Berens, S. On the conditional Rényi entropy. IEEE Trans. Inform. Theor. 2014, 60, 6801–6810. [Google Scholar] [CrossRef]
  42. Rényi, A. On measures of entropy and information. In Proceedings of the Fourth Berkeley Symposium on Mathematical Statistics and Probability, Berkeley, CA, USA, 20–30 July 1960; Volume 1. [Google Scholar]
  43. Chu, Q.; Zhang, X.; Zhu, X.; Liu, C.; Mao, L.; Ye, C.; Zhu, Q.H.; Fan, L. PlantcircBase: A database for plant circular RNAs. Mol. Plant 2017, 10, 1126–1128. [Google Scholar] [CrossRef]
  44. Manavalan, B.; Shin, T.H.; Lee, G. PVP-SVM: Sequence-based prediction of phage virion proteins using a support vector machine. Front. Microbiol. 2018, 9, 476. [Google Scholar] [CrossRef]
  45. Klapproth, C.; Sen, R.; Stadler, P.F.; Findeiß, S.; Fallmann, J. Common features in lncRNA annotation and classification: A survey. Non-Coding RNA 2021, 7, 77. [Google Scholar] [CrossRef]
  46. Lin, H.; Liang, Z.Y.; Tang, H.; Chen, W. Identifying sigma70 promoters with novel pseudo nucleotide composition. IEEE/ACM Trans. Comput. Biol. Bioinform. 2017, 16, 1316–1321. [Google Scholar] [CrossRef]
  47. Li, Q.; Zhou, W.; Wang, D.; Wang, S.; Li, Q. Prediction of anticancer peptides using a low-dimensional feature model. Front. Bioeng. Biotechnol. 2020, 8, 892. [Google Scholar] [CrossRef]
  48. Halko, N.; Martinsson, P.G.; Tropp, J.A. Finding structure with randomness: Probabilistic algorithms for constructing approximate matrix decompositions. SIAM Rev. 2011, 53, 217–288. [Google Scholar] [CrossRef]
  49. McInnes, L.; Healy, J.; Saul, N.; Großberger, L. UMAP: Uniform Manifold Approximation and Projection. J. Open Sour. Softw. 2018, 3, 861. [Google Scholar] [CrossRef]
  50. Khan, F.; Khan, M.; Iqbal, N.; Khan, S.; Muhammad Khan, D.; Khan, A.; Wei, D.Q. Prediction of Recombination Spots Using Novel Hybrid Feature Extraction Method via Deep Learning Approach. Front. Genet. 2020, 11, 1052. [Google Scholar] [CrossRef]
  51. Remita, M.A.; Halioui, A.; Malick Diouara, A.A.; Daigle, B.; Kiani, G.; Diallo, A.B. A machine learning approach for viral genome classification. BMC Bioinform. 2017, 18, 1–11. [Google Scholar] [CrossRef] [Green Version]
  52. Bonidia, R.P.; Santos, A.P.A.; de Almeida, B.L.; Stadler, P.F.; da Rocha, U.N.; Sanches, D.S.; de Carvalho, A.C. BioAutoML: Automated feature engineering and metalearning to predict noncoding RNAs in bacteria. Brief. Bioinform. 2022, 23, bbac218. [Google Scholar] [CrossRef]
  53. Randhawa, G.S.; Soltysiak, M.P.; El Roz, H.; de Souza, C.P.; Hill, K.A.; Kari, L. Machine learning using intrinsic genomic signatures for rapid classification of novel pathogens: COVID-19 case study. PLoS ONE 2020, 15, e0232391. [Google Scholar] [CrossRef] [Green Version]
  54. Naeem, S.M.; Mabrouk, M.S.; Marzouk, S.Y.; Eldosoky, M.A. A diagnostic genomic signal processing (GSP)-based system for automatic feature analysis and detection of COVID-19. Brief. Bioinform. 2021, 22, 1197–1205. [Google Scholar] [CrossRef]
  55. Arslan, H. Machine Learning Methods for COVID-19 Prediction Using Human Genomic Data. In Proceedings of the Multidisciplinary Digital Publishing Institute Proceedings, Online, 9–11 December 2020; Volume 74, p. 20. [Google Scholar]
  56. Berry, M.W. Large-scale sparse singular value computations. Int. J. Supercomput. Appl. 1992, 6, 13–49. [Google Scholar] [CrossRef]
  57. Rajamanickam, S. Efficient Algorithms for Sparse Singular Value Decomposition; University of Florida: Gainesville, FL, USA, 2009. [Google Scholar]
  58. McInnes, L.; Healy, J.; Melville, J. Umap: Uniform manifold approximation and projection for dimension reduction. arXiv 2018, arXiv:1802.03426. [Google Scholar]
  59. Becht, E.; McInnes, L.; Healy, J.; Dutertre, C.A.; Kwok, I.W.; Ng, L.G.; Ginhoux, F.; Newell, E.W. Dimensionality reduction for visualizing single-cell data using UMAP. Nat. Biotechnol. 2019, 37, 38–44. [Google Scholar] [CrossRef]
  60. Dorrity, M.W.; Saunders, L.M.; Queitsch, C.; Fields, S.; Trapnell, C. Dimensionality reduction by UMAP to visualize physical and genetic interactions. Nat. Commun. 2020, 11, 1537. [Google Scholar] [CrossRef] [Green Version]
  61. Li, M.; Si, Y.; Yang, W.; Yu, Y. ET-UMAP integration feature for ECG biometrics using Stacking. Biomed. Sig. Proc. Control 2022, 71, 103159. [Google Scholar] [CrossRef]
Figure 1. Performance analysis with five classifiers on 100 q parameters of three benchmark datasets (evaluation metric: ACC). (a) Benchmark D1—ACC; (b) Benchmark D2—ACC; (c) Benchmark D3—ACC.
Figure 1. Performance analysis with five classifiers on 100 q parameters of three benchmark datasets (evaluation metric: ACC). (a) Benchmark D1—ACC; (b) Benchmark D2—ACC; (c) Benchmark D3—ACC.
Entropy 24 01398 g001aEntropy 24 01398 g001b
Figure 2. Performance analysis with generalized entropies on 100 q parameters of three benchmark datasets (evaluation metric: ACC). (a) Benchmark D1—ACC; (b) Benchmark D2—ACC; (c) Benchmark D3—ACC.
Figure 2. Performance analysis with generalized entropies on 100 q parameters of three benchmark datasets (evaluation metric: ACC). (a) Benchmark D1—ACC; (b) Benchmark D2—ACC; (c) Benchmark D3—ACC.
Entropy 24 01398 g002
Table 1. Feature descriptors found in all studies.
Table 1. Feature descriptors found in all studies.
GroupDescriptor
Nucleic Acid CompositionNucleotide composition
Dinucleotide composition
Trinucleotide composition
Tetranucleotide composition
Pentanucleotide composition
Hexanucleotide composition
Basic k-mer
Reverse complementary k-mer
Increment in diversity
Mismatch
Subsequence
GC-content
AT/GT Ratio
Cumulative skew
kGap
Position-specific nucleotide frequency
Nucleotide content
Conformational properties
Enhanced nucleic acid composition
Composition of k-spaced Nucleic Acid Pairs
TDTopological descriptors
K-Nearest NeighborK-nearest neighbor for proteins
AutocorrelationNormalized Moreau–Broto
Moran
Geary
Dinucleotide-based auto-covariance
Dinucleotide-based cross-covariance
Dinucleotide-based auto-cross-covariance
Trinucleotide-based auto-covariance
Trinucleotide-based cross-covariance
Trinucleotide-based auto-cross-covariance
Pseudo Nucleic Acid CompositionType 1 Pseudo k-tuple nucleotide composition
Type 2 Pseudo k-tuple nucleotide composition
Pseudo k-tuple nucleotide composition
Pseudo dinucleotide composition
Numerical MappingZ-curve theory
Nucleotide Chemical Property
Accumulated Nucleotide Frequency
Electron–ion interaction pseudopotential
Pseudo electron–ion interaction pseudopotential
Binary
Orthonormal encoding
Basic one-hot
6-dimension one-hot method
Amino Acid CompositionAmino acid composition
Dipeptide composition
Tripeptide composition
Terminal end amino acid count
Amino acid pair
Secondary structure composition
Secondary structure—amino acid composition
Solvent accessibility composition
Solvent accessibility—amino acid composition
Codon composition
Protein length
Overlapping k-mers
Information-based statistics
Basic k-mer
Distance-based residue
Distance pair
Residue-Couple Model
Composition moment vector
Enhanced amino acid composition
Composition of k-spaced amino acid pairs
Dipeptide deviation from expected mean
Grouped amino acid composition
Enhanced grouped amino acid composition
Composition of k-spaced amino acid group pairs
Grouped dipeptide composition
Grouped tripeptide composition
kGap
Position-specific nucleotide frequency
Pseudo-Amino Acid CompositionType 1 PseAAC
Type 2 PseAAC
Dipeptide (or Type 3) PseAAC
General parallel correlation PseAAC
General series correlation PseAAC
Pseudo k-tuple reduced AAC (type 1 to type 16)
CTDComposition
Transition
Distribution
Sequence-OrderSequence-order-coupling number
Quasi-sequence-order
Profile-based FeaturesSignal average
Signal peak area
PSSM (Position-Specific Scoring Matrix) profile
Profile-based physicochemical distance
Distance-based top-n-gram
Top-n-gram
Sequence conservation score
Frequency profile matrix
Conjoint TriadConjoint Triad
Conjoint k-spaced triad
Proteochemometric DescriptorsPrincipal component analysis
Principal component analysis (2D and 3D)
Factor analysis
Factor analysis (2D and 3D)
Multidimensional scaling
Multidimensional scaling (2D and 3D)
BLOSUM and PAM matrix-derived
Biophysical quantitative properties
Amino acid properties
Molecular descriptors
Sequence SimilarityGene Ontology (GO) similarity
Sequence Alignment
BLAST matrix
Structure CompositionSecondary structure
Solvent accessible surface area
Secondary structure binary
Disorder
Disorder content
Disorder binary
Torsional angles
DNA shape features
Physicochemical PropertyAAindex
Z-scale
Physicochemical n-Grams
Dinucleotide physicochemical
Trinucleotide physicochemical
Table 2. The best and worst parameter (q) of each benchmark dataset and classifier, taking into account the ACC metric.
Table 2. The best and worst parameter (q) of each benchmark dataset and classifier, taking into account the ACC metric.
DatasetGaussianNBRFBaggingMLPCatBoost
qACCqACCqACCqACCqACC
D12.70.93700.40.94302.70.94002.20.93802.30.9440
9.20.47609.60.73609.60.727010.00.50609.60.747
D21.50.79805.30.82205.70.80800.90.78004.00.8300
9.60.521010.00.651010.00.61709.90.50609.20.6800
D38.70.70087.80.69102.00.71571.50.71841.10.7282
1.30.60629.80.59859.50.59620.10.68605.70.6610
Table 3. Performance with different entropic index (q) values for the sigma70 promoter classification problem.
Table 3. Performance with different entropic index (q) values for the sigma70 promoter classification problem.
DatasetqClassifierACCRecallF1 ScoreAUCBACC
D40.5RF0.65940.25560.34230.62790.5647
CatBoost0.65630.19730.28480.62330.5487
2.0RF0.66870.30940.39320.61080.5845
CatBoost0.66410.20630.29870.63010.5567
3.0RF0.66720.30490.38860.61500.5822
CatBoost0.66250.23770.32820.63190.5629
4.0RF0.66410.28250.36840.61630.5746
CatBoost0.66560.24660.33850.64150.5674
5.0RF0.66410.28250.36840.63480.5746
CatBoost0.67340.26460.35980.63750.5775
Table 4. Performance with different entropic index (q) values for the anticancer peptide classification problem.
Table 4. Performance with different entropic index (q) values for the anticancer peptide classification problem.
DatasetqClassifierACCRecallF1 ScoreAUCBACC
D50.5RF0.70190.59520.61730.74370.6847
CatBoost0.69230.38100.50000.74880.6421
2.0RF0.70190.54760.59740.74540.6770
CatBoost0.65380.42860.50000.75000.6175
3.0RF0.72120.57140.62340.77480.6970
CatBoost0.68270.42860.52170.73850.6417
4.0RF0.70190.52380.58670.78230.6732
CatBoost0.69230.47620.55560.76420.6575
5.0RF0.72110.54760.61330.78130.6932
CatBoost0.69230.47620.55560.76000.6575
Table 5. Performance with different entropic index (q) values for the SARS-CoV-2 (COVID-19) classification problem.
Table 5. Performance with different entropic index (q) values for the SARS-CoV-2 (COVID-19) classification problem.
DatasetqClassifierACCRecallF1 ScoreAUCBACC
D60.5RF0.99890.99920.99941.00000.9985
CatBoost0.99821.00000.99900.99990.9947
2.0RF0.99961.00000.99981.00000.9990
CatBoost0.99510.99960.99711.00000.9862
3.0RF1.00001.00001.00001.00001.0000
CatBoost0.99961.00000.99981.00000.9990
4.0RF1.00001.00001.00001.00001.0000
CatBoost0.99961.00000.99981.00000.9990
5.0RF1.00001.00001.00001.00001.0000
CatBoost1.00001.00001.00001.00001.0000
Table 6. Performance of the proposed approach (Tsallis) vs. Shannon entropy (best results in bold). A tie counts one win for each approach.
Table 6. Performance of the proposed approach (Tsallis) vs. Shannon entropy (best results in bold). A tie counts one win for each approach.
DatasetClassifierEntropyqACCRecallF1 ScoreBACC
D1CatBoostTsallis2.30.94200.96730.94370.9421
Shannon-0.94200.96510.94350.9421
D2CatBoostTsallis4.00.81400.77600.80530.8153
Shannon-0.80800.75820.79700.8115
D3CatBoostTsallis1.10.72310.38690.47240.6342
Shannon-0.72070.38860.47080.6334
D4RFTsallis2.00.66870.30940.39320.5845
Shannon-0.65630.25560.34030.5623
D5RFTsallis3.00.72120.57140.62340.6970
Shannon-0.71150.54760.60530.6851
D6RFTsallis5.00.99840.98460.99150.9922
Shannon-0.99850.98880.99220.9942
Mean-Tsallis-0.81120.66590.70490.7776
Shannon-0.80610.65070.69150.7714
Gain---0.51%1.52%1.34%0.62%
Wins-Tsallis-5455
Shannon-2212
Table 7. Performance of the proposed approach (Tsallis) vs. SVD vs. UMAP. A tie counts one win for each approach.
Table 7. Performance of the proposed approach (Tsallis) vs. SVD vs. UMAP. A tie counts one win for each approach.
DatasetReductionACCRecallF1 ScoreBACC
D1Tsallis (q = 2.3)0.94300.96500.94380.9434
SVD0.49800.00000.00000.4982
UMAP0.49800.99630.66320.4981
D2Tsallis (q = 4.0)0.81200.77180.80300.8114
SVD0.50040.00160.00320.5008
UMAP0.49940.00000.00000.5000
D3Tsallis (q = 1.1)0.73070.35380.45410.6310
SVD0.53890.71320.49420.5834
UMAP0.31910.99330.48250.4967
D5Tsallis (q = 3.0)0.67200.51810.55150.6508
SVD0.74030.76300.77520.7261
UMAP0.40210.00000.00000.5000
D7Tsallis (q = 3.0)0.73710.67110.69470.7337
SVD0.54380.00000.00000.4992
UMAP0.51430.18240.11470.4963
D8Tsallis (q = 1.1)0.65000.61110.62770.6525
SVD0.80230.85750.78430.8171
UMAP0.63260.77280.65440.6511
D9Tsallis (q = 9.2)0.94890.94810.95070.9481
SVD0.55860.64330.55170.6433
UMAP0.59920.65280.61670.6528
MeanTsallis0.78480.69130.71790.7673
SVD0.59750.42550.37270.6097
UMAP0.49500.51390.36160.5421
WinsTsallis5345
SVD2232
UMAP0200
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Bonidia, R.P.; Avila Santos, A.P.; de Almeida, B.L.S.; Stadler, P.F.; Nunes da Rocha, U.; Sanches, D.S.; de Carvalho, A.C.P.L.F. Information Theory for Biological Sequence Classification: A Novel Feature Extraction Technique Based on Tsallis Entropy. Entropy 2022, 24, 1398. https://doi.org/10.3390/e24101398

AMA Style

Bonidia RP, Avila Santos AP, de Almeida BLS, Stadler PF, Nunes da Rocha U, Sanches DS, de Carvalho ACPLF. Information Theory for Biological Sequence Classification: A Novel Feature Extraction Technique Based on Tsallis Entropy. Entropy. 2022; 24(10):1398. https://doi.org/10.3390/e24101398

Chicago/Turabian Style

Bonidia, Robson P., Anderson P. Avila Santos, Breno L. S. de Almeida, Peter F. Stadler, Ulisses Nunes da Rocha, Danilo S. Sanches, and André C. P. L. F. de Carvalho. 2022. "Information Theory for Biological Sequence Classification: A Novel Feature Extraction Technique Based on Tsallis Entropy" Entropy 24, no. 10: 1398. https://doi.org/10.3390/e24101398

APA Style

Bonidia, R. P., Avila Santos, A. P., de Almeida, B. L. S., Stadler, P. F., Nunes da Rocha, U., Sanches, D. S., & de Carvalho, A. C. P. L. F. (2022). Information Theory for Biological Sequence Classification: A Novel Feature Extraction Technique Based on Tsallis Entropy. Entropy, 24(10), 1398. https://doi.org/10.3390/e24101398

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop