Search tips
Search criteria 


Logo of peerjLatest ArticlesFor AuthorsEditorial BoardPeerJPeerJ
PeerJ. 2016; 4: e2135.
Published online 2016 June 21. doi:  10.7717/peerj.2135
PMCID: PMC4924126

The impact of feature selection on one and two-class classification performance for plant microRNAs

Academic Editor: Eiji Nambara


MicroRNAs (miRNAs) are short nucleotide sequences that form a typical hairpin structure which is recognized by a complex enzyme machinery. It ultimately leads to the incorporation of 18–24 nt long mature miRNAs into RISC where they act as recognition keys to aid in regulation of target mRNAs. It is involved to determine miRNAs experimentally and, therefore, machine learning is used to complement such endeavors. The success of machine learning mostly depends on proper input data and appropriate features for parameterization of the data. Although, in general, two-class classification (TCC) is used in the field; because negative examples are hard to come by, one-class classification (OCC) has been tried for pre-miRNA detection. Since both positive and negative examples are currently somewhat limited, feature selection can prove to be vital for furthering the field of pre-miRNA detection. In this study, we compare the performance of OCC and TCC using eight feature selection methods and seven different plant species providing positive pre-miRNA examples. Feature selection was very successful for OCC where the best feature selection method achieved an average accuracy of 95.6%, thereby being ~29% better than the worst method which achieved 66.9% accuracy. While the performance is comparable to TCC, which performs up to 3% better than OCC, TCC is much less affected by feature selection and its largest performance gap is ~13% which only occurs for two of the feature selection methodologies. We conclude that feature selection is crucially important for OCC and that it can perform on par with TCC given the proper set of features.

Keywords: MicroRNA, Machine learning, Feature selection, Plant, One-class classification, Two-class classification


Gene regulation is of prime importance in all living organisms and there are multiple levels at which gene expression can be modulated. MicroRNAs (miRNAs) play a role in post-transcriptional gene regulation (Erson-Bensan, 2014) and, among other functions, fine-tune the amount of translated protein product (Saçar & Allmer, 2013b). Mature miRNAs are short nucleotide sequences discovered about two decades ago (Lee, Feinbaum & Ambros, 1993). From databases which host miRNAs like miRBase (Kozomara & Griffiths-Jones, 2011) it can be gleaned that miRNAs exist in a wide range of organisms ranging from viruses (Grey, 2015) to plants (Yousef, Allmer & Khalifa, 2016). It has also been proposed that the plant miRNA system may have evolved independently (Chapman & Carrington, 2007) and some organisms like yeasts also display differences to the canonical pathway (Ender & Meister, 2010). Any regulatory element itself may be miss-regulated and miRNAs are no exception, and therefore, have been implicated in, for example, human diseases (Alural et al., 2014; Alural et al., 2015) and in plant stress response (Zhang et al., 2010). While miRNAs may lead to inter-kingdom communication in special cases (Bağcı & Allmer, 2012), it is not likely that there is extensive communication among eukaryotes (Bağcı & Allmer, 2016). Experimentally detected and/or validated miRNAs are available in databases such as miRBase (Griffiths-Jones et al., 2008) and miRTarBase (Hsu et al., 2011). MicroRNAs’ effect can only be established when it is co-expressed with its targets (Saçar & Allmer, 2013b), which complicates experimental analysis since only a fraction of the genome is expressed at a given time, in a tissue, or in response to stress conditions; and testing all conditions experimentally is elusive. Additionally, such an analysis needs to be performed on transcript and protein level, concurrently, over multiple time points to establish a causative relationship. Therefore, it seems impossible to experimentally detect all possible miRNAs of any higher eukaryotic organism (Yousef et al., 2008; Ding, Zhou & Guan, 2010; Wu et al., 2011; Ritchie, Gao & Rasko, 2012). Moreover, it has become clear that even among the experimentally validated miRNAs in miRBase and mirTarBase, there may be dubious examples (Saçar, Hamzeiy & Allmer 2013). Therefore, carefully designed computational experiments are required to complement experimental approaches for miRNA detection.

Many computational approaches to miRNA detection have been proposed and most of them derive numerical features (Sacar & Allmer, 2013) to describe a pre-miRNA and then use machine learning to establish a model for miRNA identification (Allmer & Yousef, 2012; Yousef, Allmer & Khalifa, 2015; Saçar & Allmer, 2013a; Saçar & Allmer, 2014; Allmer, 2014). Of these approaches, most, with few exceptions (Yousef et al., 2008; Yousef, Allmer & Khalifa, 2015; Koski et al., 2005), employ two class classification; the latter has been compared previously (Saçar & Allmer, 2013a; De On Lopes, Schliep & De Lf de Carvalho, 2014). Classification in machine learning depends on positive examples for training the classifier in case of one-class classification (OCC) and additionally on negative data in case of two-class classification (TCC). The negative data, however, proves to be difficult to establish (if not impossible), so that all negative datasets currently in use are based on arbitrary selection of examples from parts of a genome deemed not miRNA genic or from randomly generated sequences. While both approaches are questionable, they present the only alternative to using OCC and in the absence of proper benchmark data need to be used for TCC (Allmer, 2012). Since OCC only needs examples for the target class (here positives), it can obliterate the need to define artificial negative examples (Manevitz & Yousef, 2002; Manevitz & Yousef, 2007) and can be used to differentiate between target and unknown class. We have recently analyzed the use of OCC for miRNA detection in plants and found that it was competitive in comparison to TCC although the analysis was unduly biased towards TCC (Yousef, Allmer & Khalifa, 2016). Our previous study also showed that among the hundreds of features proposed for miRNA parameterization (Sacar & Allmer, 2013) some are more discriminative than others. Since feature selection is NP-hard (Amaldi & Kann, 1998), selecting the best subset from more than 1,000 features on a per dataset basis is not achievable. Feature selection has been investigated before, but mostly for TCC (Paul, Magdon-Ismail & Drineas, 2015; Guyon et al., 2002; Ahsen et al., 2012), while only little has been done for OCC (Lorena, Carvalho & Lorena, 2015; Xuan et al., 2011a; Hall et al., 2009). In this study, we used different feature selection approaches and compared their effectiveness for OCC and TCC classification performance.

Both machine learning approaches, OCC and TCC, benefit from feature selection. While feature selection is essential for OCC and a difference of about 30% accuracy can be observed, the maximum difference for TCC is ~10%. Moreover, for TCC 7 out of 8 feature selection methods lead to accuracy greater than 90% whereas such high accuracy was only achieved for two methods when using OCC. For the LIG feature selection method, intended as a negative control, both classifiers display lowest performance but TCC is about 20% better than OCC. With increasing accuracy (i.e., better feature selection for OCC), the accuracy for TCC also increases; except for the SFC which is best for OCC but only third best for TCC. While the performance difference for LIG is large, it decreases with the use of better feature selection methods. TCC is only 3% better when the SFC feature selection method is considered, which provided the best performance for OCC. A difference in performance among plant species was observed for both classifiers, but for TCC it was about 5% whereas for OCC it was 15%. In conclusion, feature selection is essential for OCC, but does not affect TCC as much. We propose that due to the lack of true negative data, more focus should be put on the further development of OCC approaches to pre-miRNA detection.

Materials and Methods


Positive examples for pre-miRNAs from selected plant species were downloaded from miRBase (Griffiths-Jones et al., 2008) (Releases 20 and 21). Glycine max (gma), Zea mays (zma), Sorghum bicolor (sbi), Physcomitrella patens (ppt), Arabidopsis thaliana (ath), Populus trichocarpa (ptc), and Oryza sativa (osa) make up the positive dataset. Negative examples for miRNAs consisted of 980 pseudo pre-miRNAs from the PlantMiRNAPred dataset (Xuan et al., 2011b). For these data, all pre-miRNA features were calculated as described previously (Sacar & Allmer, 2013; Yousef, Allmer & Khalifa, 2015; Saçar, Bağcı & Allmer, 2014). We chose plant pre-miRNAs with large amount of pre-miRNA examples and from different clades for this study. Additionally, plant miRNAs have not been investigated as extensively as metazoan miRNAs which adds to the reason to choose plant pre-miRNAs.

One class classification

For one-class classification the DDtools (Tax, 2015) implementation of an OCC was utilized. A 100-fold Monte Carlo cross validation (Xu & Liang, 2001) was performed using randomly sampled 90% of the positive data for training and 10% for testing. Moreover, the pseudo negative sequences were injected as unknown class during testing. We employed k-means in this study as previously described (Yousef, Allmer & Khalifa, 2016) since it performed well in respect to OCC although it is a clustering algorithm. During learning, labeled examples are clustered (miRNAs and unknown) and during testing and in prediction, the label of the closest cluster is assigned to the sample.

Two class classification

Support Vector Machines (SVMs) are used for machine learning and were first proposed by Vapnik (1995). In bioinformatics and in the field of pre-miRNA detection, SVMs have been used (Ding, Zhou & Guan, 2010; Wu et al., 2011; Xuan et al., 2011b; Ng & Mishra, 2007). Here, the WEKA library (Gewehr, Szugat & Zimmer, 2007) SVM implementation which is based on LibSVM (Chang & Lin, 2011) was utilized. The radial basis function was set to a gamma value of 0.7 and the cost parameter was chosen to be 4.0 and the normalization option was set to true. Any machine learning algorithm needs initial training and we performed a 10 fold Monte Carlo cross validation (Xu & Liang, 2001) during learning, by employing random sampling using 90% of the data for training and 10% for testing.

Feature selection strategies

Feature selection has been shown to be an NP-hard problem and, therefore, other approximate feature selection strategies are being developed. In machine learning for pre-miRNAs, more than 1,000 features have been proposed which makes feature selection especially hard. To investigate the impact of feature selection on model performance for OCC and TCC, four negative and four positive feature selection methods were designed. Previously, we found that a set of 50–100 features may be sufficient for successful pre-miRNA detection (Sacar & Allmer, 2013). Using more than 50 features increases the likelihood that the feature set contains some features which may conceal differences among feature selection methods. Therefore, a feature set size of 50 was selected for model training in this study.

We have previously performed feature selection for OCC (Yousef et al., 2016) using similar feature selection methods as we propose here, but it is important to compare the impact between OCC and TCC.

Eight feature selection methods were devised and four of them were expected to lead to low performance while the remaining methods were thought to perform well. The former were selecting features with low information gain (LIG), random feature selection (RFS), selecting random feature from feature clusters (RFC), and selecting features from clusters (SFC). The latter were selecting features with high information gain (HIG), selecting the highest information gain from feature clusters (HIC), zero-norm feature selection (ZNF), and Pearson correlation-based feature selection (PCF).

All feature selection methods except for the last two were performed using KNIME (Berthold et al., 2009) and the selected features are available in Table S1; information on how to calculate them are provided in File S1. The workflows for our feature selection methods, developed in KNIME, are available for download from our website:

In order to calculate LIG and HIG, for each dataset, the information gain (IG) among features was established (using KNIME’s InformationGainCalculator node) and the 50 features with lowest IG (LIG) or highest IG (HIG) were selected. For RFS, 50 random features were selected using the Row Sampling node in KNIME. To establish RFC, features were clustered using WEKA k-Means implementation in KNIME (k = 100). From each cluster a random feature was selected and from the 100 random features the final set of 50 features was selected randomly (KNIME’s Row Sampling approach). For SFC, clustering was performed as for RFC. Clusters were ordered by number of cluster members (largest to smallest) and the 50 features were chosen from the top. To derive HIC, the same clustering approach as for RFC and SFC was taken, but the features in each cluster were ranked according to IG and the best one was selected. The selected 100 features were again ranked using IG and the best 50 were selected. ZNF is defined to be the non-zero values for all feature vectors of positive examples. Among the non-zero ones, the 50 features with highest sum of values were selected. PCF was established according to Lorena, Carvalho & Lorena (2015) and after Pearson clustering the features with lowest correlation score were retained. Feature selection was performed on a per species basis which led to the selection of different features (Table S1). Combined feature selection uses the occurrence of selected features among the seven selected species and five mixed datasets in respect to the top 100 features. The Features were ranked according to their frequency and top 50 were selected (Table S1).

Results and Discussion

Eight feature selection methods were designed and they were applied to seven plant datasets. For each dataset OCC (100) and TCC (10) models were established using Monte Carlo cross validation (MCCV). Feature selection was performed on a per plant dataset basis. The 50 features selected varied to some extend and, therefore, we defined another feature set (indicated by ‘comb’) which was created by selecting the features ordered by decreasing incidence based on the individual selections. The selected features are provided in Table S1 by their acronyms which are explained in more detailed in our previous studies (Sacar & Allmer, 2013; Saçar & Allmer, 2013a).

We applied the eight feature selection methods to the seven plant species’ datasets individually and recorded the model performance. Figure 1 shows the average model performance (OCC: 100, TCC: 10 fold cross validation) for the best feature selection method we found (SFC) and the worst one (LIG).

Figure 1
Best (SFC) versus worst (LIG) feature selection method on per species feature selection.

Sensitivity was the performance measure most affected for both machine learning approaches (Fig. 1). For TCC the average accuracy among plant species dropped about 10% between SFC and LIG while it dropped about 30% for OCC. The results for the remaining six feature selection methods are presented in Table S2.

The impact on using combined feature selection for SFC and LIG is quite similar to individual feature selection (Fig. 2). The combined features were not calculated for PCF and ZNF since combination of features was not supported by our workflow in this case. Overall accuracy is slightly reduced for the combined feature selection by on average 1% (OCC) and 2% (TCC) when compared to individual feature selection.

Figure 2
Best (SFC) versus worst (LIG) feature selection method on consensus feature selection.

The performance analysis of the remaining feature selection methods are presented in Table S2. In order to compare the performance of all feature selection methods for the two machine learning approaches, the average model accuracy was plotted (Fig. 3). It is striking that for most (six out of eight) TCC performance results the accuracy is above 95% for all plant species.

Figure 3
Model accuracy comparison between OCC and TCC in respect to feature selection method.

For OCC, the performance is best for SFC where, on average, for plant species it achieves more than 95% accuracy (Fig. 3). All other feature selection methods do not lead to high performing models with HIC being the second best, followed by RFS and PCF. For most feature selection methodologies the accuracy among plant species is quite similar for TCC, but for OCC the differences are much larger.

In order to compare the variance in performance between OCC and TCC, the difference between TCC and OCC accuracy was calculated (TCCACC–OCCACC) and is presented in Fig. 4. Positive values signify better performance of TCC.

Figure 4
Comparison of the effect of feature selection on two-class versus one-class classification.

The most accurate OCC model is on the left (SFC) and it is seen that TCC is outperformed by OCC on several plant species (ath, gma, osa, and ppt). Figure 4 shows that OCC is more affected by feature selection than TCC and, therefore, with increasing effectiveness of the feature selection methodology, the difference between classifiers diminishes. For improper feature selection it can reach up to about 30%, whereas it drops to almost similar performance for the best feature selection method in this study (SFC, ~0.6% on average).


Many general purpose feature selection methods have been described or used in bioinformatics (Saeys, Inza & Larrañaga, 2007). For OCC feature selection nothing has been done in the area of pre-miRNA detection while one study investigated feature selection based on OCC for mature miRNA prediction (Xuan et al., 2011a). When considering two class classification of pre-miRNAs SVM recursive feature elimination (RFE) has been used (Shu et al., 2015). Meng et al., (2014) also used RFE, but modified it and compared to principal component analysis (PCA), correlation-based feature subset selection (CFS), and not using any filtering. They report the best accuracy for SVM using their back SVM-RFE FS with 97.2% closely followed by PCA using SVM with 97.0% accuracy. One approach used genetic algorithm in combination with information gain and also taking into account feature redundancy for FS and achieved almost 99.5% accuracy, alas on a limited dataset (Xuan et al., 2011c). These competing methods using different strategies for FS in pre-miRNA detection do not refer to OCC. However, they clearly show that feature selection has a large impact on model performance. The previous methodologies used correlation among features or feature redundancy for FS but did not put a clear focus on the correlation issue. We, therefore, devised eight feature selection methodologies with a focus on feature correlation and applied them to several plant miRNA datasets. Feature selection was performed on a per plant species basis, but we also investigated the combined feature set using the features shared among species; both of which were not done in previous approaches. Our SFC feature selection methodology was particularly successful and there was no great difference for feature selection on a per plant basis or when combined (Figs. 1A and and2A).2A). As expected, the LIG methodology did not perform well at all and was intended as a negative control. However, the SVM learner was not nearly as much affected as the OCC one (Figs. 1B and and2B)2B) although sensitivity was strongly affected for both learners.

Of the eight feature selection methods tested in this study, only 3 show good performance for OCC (SFC, HIC, and RFS; Fig. 3) while only two did not seem applicable for SVM (LIG and PCF; Fig. 3). For most feature selection methods average SVM performance is above 95% while OCC performance is generally below 90% (Fig. 3).

It is instructive to analyze the performance difference between SVM and OCC. Figure 4 shows the performance difference and it can be seen that for most feature selection methods SVM performs better than OCC (positive values; Fig. 3). However, the SFC feature selection method which is among the best for SVM clearly performs best for OCC and the latter can surpass the SVM performance for several of the selected plant species.

From this study it can be concluded that the more successful the feature selection the less difference between OCC and TCC model performance and the better the overall model performance. Thus we conclude, that in the absence of missing negative data OCC should be used and, therefore, additional feature selection strategies should be tried to improve its performance.

Supplemental Information


File S1

Work Feature Selection Workflows:

Table S1

Supplementary Table 1:

The results for the remaining six feature selection methods.


Table S2

Supplementary Table 2:

All feature selection methods except for the last two and the selected features as well as information on how to calculate them.

Funding Statement

The work was supported by the Scientific and Technological Research Council of Turkey (grant number 113E326) to JA. The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.

The following grant information was disclosed by the authors:

Scientific and Technological Research Council 113E326.

Additional Information and Declarations

Competing Interests

The authors declare there are no competing interests.

Author Contributions

Waleed Khalifa conceived and designed the experiments, performed the experiments, prepared figures and/or tables, reviewed drafts of the paper.

Malik Yousef and Jens Allmer conceived and designed the experiments, performed the experiments, analyzed the data, contributed reagents/materials/analysis tools, wrote the paper, prepared figures and/or tables, reviewed drafts of the paper.

Müşerref Duygu Saçar Demirci performed the experiments, prepared figures and/or tables, reviewed drafts of the paper.

Data Availability

The following information was supplied regarding data availability:

All data has been supplied as Supplemental Information.


Ahsen et al. (2012) Ahsen ME, Singh NK, Boren T, Vidyasagar M, White MA. 2012 IEEE 51st IEEE conference on decision and control (CDC) 2012. A new feature selection algorithm for two-class classification problems and application to endometrial cancer; pp. 2976–2982.
Allmer (2012) Allmer J. A call for benchmark data in mass spectrometry-based proteomics. Journal of Integrated OMICS. 2012;2(2) doi: 10.5584/jiomics.v2i2.113. Epub ahead of print Oct 28 2012. [Cross Ref]
Allmer (2014) Allmer J. Computational and bioinformatics methods for microRNA gene prediction. Methods in Molecular Biology. 2014;1107:157–175. doi: 10.1007/978-1-62703-748-8_9. [PubMed] [Cross Ref]
Allmer & Yousef (2012) Allmer J, Yousef M. Computational methods for ab initio detection of microRNAs. Frontiers in Genetics. 2012;3:209. [PMC free article] [PubMed]
Alural et al. (2014) Alural B, Duran GA, Tufekci KU, Allmer J, Onkal Z, Tunali D, Genc K, Genc S. Epo mediates neurotrophic, neuroprotective, anti-oxidant, and anti-apoptotic effects via downregulation of mir-451 and mir-885-5p in SH-SY5Y neuron-like cells. Frontiers in Immunology. 2014;5(September):475. [PMC free article] [PubMed]
Alural et al. (2015) Alural B, Ozerdem A, Allmer J, Genc K, Genc S. Lithium protects against paraquat neurotoxicity by NRF2 activation and miR-34a inhibition in SH-SY5Y cells. Frontiers in Cellular Neuroscience. 2015;9:209. [PMC free article] [PubMed]
Amaldi & Kann (1998) Amaldi E, Kann V. On the approximability of minimizing nonzero variables or unsatisfied relations in linear systems. Theoretical Computer Science. 1998;209(1–2):237–260. doi: 10.1016/S0304-3975(97)00115-1. [Cross Ref]
Bağcı & Allmer (2012) Bağcı C, Allmer J. 2012 7th international symposium on health informatics and bioinformatics. 2012. Removing contamination from genomic sequences based on vector reference libraries; pp. 118–122.
Bağcı & Allmer (2016) Bağcı C, Allmer J. One step forward, two steps back; xeno-microRNAs reported in breast milk are artifacts. PLoS ONE. 2016;11(1):e2135 doi: 10.1371/journal.pone.0145065. [PMC free article] [PubMed] [Cross Ref]
Berthold et al. (2009) Berthold MR, Cebron N, Dill F, Gabriel TR, Kötter T, Meinl T, Ohl P, Thiel K, Wiswedel B. KNIME—the Konstanz Information Miner. ACM SIGKDD Explorations Newsletter. 2009;11(1):26–31. doi: 10.1145/1656274.1656280. [Cross Ref]
Chang & lin (2011) Chang C-C, Lin C-J. LIBSVM. ACM Transactions on Intelligent Systems and Technology. 2011;2(3):1–27.
Chapman & Carrington (2007) Chapman EJ, Carrington JC. Specialization and evolution of endogenous small RNA pathways. Nature Reviews Genetics. 2007;8(11):884–896. doi: 10.1038/nrg2179. [PubMed] [Cross Ref]
De On Lopes, Schliep & De Lf de Carvalho (2014) De On Lopes I, Schliep A, De Lf de Carvalho AC. The discriminant power of RNA features for pre-miRNA recognition. BMC Bioinformatics. 2014;15(1):124. doi: 10.1186/1471-2105-15-124. [PMC free article] [PubMed] [Cross Ref]
Ding, Zhou & Guan (2010) Ding J, Zhou S, Guan J. MiRenSVM: towards better prediction of microRNA precursors using an ensemble SVM classifier with multi-loop features. BMC Bioinformatics. 2010;11:S11. doi: 10.1186/1471-2105-11-S11-S11. [PMC free article] [PubMed] [Cross Ref]
Ender & Meister (2010) Ender C, Meister G. Argonaute proteins at a glance. Journal of Cell Science. 2010;123(11):1819–1823. doi: 10.1242/jcs.055210. [PubMed] [Cross Ref]
Erson-Bensan (2014) Erson-Bensan AE. Introduction to microRNAs in biological systems. Methods in Molecular Biology. 2014;1107:1–14. doi: 10.1007/978-1-62703-748-8_1. [PubMed] [Cross Ref]
Gewehr, Szugat & Zimmer (2007) Gewehr JE, Szugat M, Zimmer R. BioWeka–extending the Weka framework for bioinformatics. Bioinformatics. 2007;23(5):651–653. doi: 10.1093/bioinformatics/btl671. [PubMed] [Cross Ref]
Grey (2015) Grey F. Role of microRNAs in herpesvirus latency and persistence. Journal of General Virology. 2015;96(Pt 4):739–751. doi: 10.1099/vir.0.070862-0. [PubMed] [Cross Ref]
Griffiths-Jones et al. (2008) Griffiths-Jones S, Saini HK, Van Dongen S, Enright AJ. miRBase: tools for microRNA genomics. Nucleic Acids Research. 2008;36(Database issue):D154–D158. doi: 10.1093/nar/gkn221. [PMC free article] [PubMed] [Cross Ref]
Guyon et al. (2002) Guyon I, Weston J, Barnhill S, Vapnik V. Gene selection for cancer classification using support vector machines. Machine Learning. 2002;46:389–422. doi: 10.1023/A:1012487302797. [Cross Ref]
Hall et al. (2009) Hall M, Frank E, Holmes G, Pfahringer B, Reutemann P, Witten IH. The WEKA data mining software. ACM SIGKDD Explorations Newsletter. 2009;11:10–18. doi: 10.1145/1656274.1656278. [Cross Ref]
Hsu et al. (2011) Hsu S-D, Lin F-M, Wu W-Y, Liang C, Huang W-C, Chan W-L, Tsai W-T, Chen G-Z, Lee C-J, Chiu C-M, Chien C-H, Wu M-C, Huang C-Y, Tsou A-P, Huang H-D. miRTarBase: a database curates experimentally validated microRNA-target interactions. Nucleic Acids Research. 2011;39(Database issue):D163–D169. doi: 10.1093/nar/gkq1107. [PMC free article] [PubMed] [Cross Ref]
Koski et al. (2005) Koski LB, Gray MW, Lang BF, Burger G. AutoFact: an automatic functional annotation and classification tool. BMC Bioinformatics. 2005;6:151. doi: 10.1186/1471-2105-6-151. [PMC free article] [PubMed] [Cross Ref]
Kozomara & Griffiths Jones (2011) Kozomara A, Griffiths-Jones S. miRBase: integrating microRNA annotation and deep-sequencing data. Nucleic Acids Research. 2011;39(Database issue):D152–D157. doi: 10.1093/nar/gkq1027. [PMC free article] [PubMed] [Cross Ref]
Lee, Feinbaum & Ambros (1993) Lee RC, Feinbaum RL, Ambros V. The C elegans heterochronic gene lin-4 encodes small RNAs with antisense complementarity to lin-14. Cell. 1993;75(5):843–854. doi: 10.1016/0092-8674(93)90529-Y. [PubMed] [Cross Ref]
Lorena, Carvalho & Lorena (2015) Lorena LHN, Carvalho ACPLF, Lorena AC. Filter feature selection for one-class classification. Journal of Intelligent and Robotic Systems. 2015;80:227–243. doi: 10.1007/s10846-014-0101-2. [Cross Ref]
Manevitz & Yousef (2002) Manevitz LM, Yousef M. One-class SVMs for document classification. Journal of Machine Learning Research. 2002;2:139–154.
Manevitz & Yousef (2007) Manevitz L, Yousef M. One-class document classification via neural networks. Neurocomputing. 2007;70(7–9):1466–1481. doi: 10.1016/j.neucom.2006.05.013. [Cross Ref]
Meng et al. (2014) Meng J, Liu D, Sun C, Luan Y. Prediction of plant pre-microRNAs and their microRNAs in genome-scale sequences using structure-sequence features and support vector machine. BMC Bioinformatics. 2014;15:423. doi: 10.1186/s12859-014-0423-x. [PMC free article] [PubMed] [Cross Ref]
Ng & Mishra (2007) Ng KLS, Mishra SK. De novo SVM classification of precursor microRNAs from genomic pseudo hairpins using global and intrinsic folding measures. Bioinformatics. 2007;23(11):1321–1330. doi: 10.1093/bioinformatics/btm026. [PubMed] [Cross Ref]
Paul, Magdon-Ismail & Drineas (2015) Paul S, Magdon-Ismail M, Drineas P. Feature selection for linear SVM with provable guarantees. Journal of Machine Learning Research. 2015;38:735–743.
Ritchie, Gao & Rasko (2012) Ritchie W, Gao D, Rasko JEJ. Defining and providing robust controls for microRNA prediction. Bioinformatics. 2012;28(8):1058–1061. doi: 10.1093/bioinformatics/bts114. [PubMed] [Cross Ref]
Sacar & Allmer (2013) Sacar MD, Allmer J. 2013 8th international symposium on health informatics and bioinformatics. 2013. Data mining for microrna gene prediction: on the impact of class imbalance and feature number for microrna gene prediction; pp. 1–6.
Saçar & Allmer (2013a) Saçar MD, Allmer J. Proceedings of the international conference on bioinformatics models, methods and algorithms. 2013a. Comparison of four ab initio microrna prediction tools; pp. 190–195.
Saçar & Allmer (2013b) Saçar MD, Allmer J. Current limitations for computational analysis of miRNAs in cancer. Pakistan Journal of Clinical and Biomedical Research. 2013b;1(2):3–5.
Saçar & Allmer (2014) Saçar MD, Allmer J. Machine learning methods for microRNA gene prediction. Methods in Molecular Biology. 2014;1107:177–187. doi: 10.1007/978-1-62703-748-8_10. [PubMed] [Cross Ref]
Saçar, Bağcı & Allmer (2014) Saçar MD, Bağcı C, Allmer J. Computational prediction of microRNAs from Toxoplasma gondii potentially regulating the hosts’ gene expression. Genomics, Proteomics Bioinformatics. 2014;12(5):228–238. doi: 10.1016/j.gpb.2014.09.002. [PMC free article] [PubMed] [Cross Ref]
Saçar, Hamzeiy & Allmer (2013) Saçar MD, Hamzeiy H, Allmer J. Can MIRBase provide positive data for machine learning for the detection of miRNA hairpins? Journal of Integrative Bioinformatics. 2013;10(2):215. [PubMed]
Saeys, Inza & Larrañaga (2007) Saeys Y, Inza I, Larrañaga P. A review of feature selection techniques in bioinformatics. Bioinformatics. 2007;23(19):2507–2517. doi: 10.1093/bioinformatics/btm344. [PubMed] [Cross Ref]
Shu et al. (2015) Shu J, Chiang K, Zempleni J, Cui J. Computational characterization of exogenous microRNAs that can be transferred into human circulation. PLoS ONE. 2015;10(11):e2135 doi: 10.1371/journal.pone.0140587. [PMC free article] [PubMed] [Cross Ref]
Tax (2015) Tax DMJ. DDtools, the data description toolbox for Matlab. 2015.
Vapnik (1995) Vapnik VN. The nature of statistical learning theory. Springer-Verlag; New York: 1995.
Wu et al. (2011) Wu Y, Wei B, Liu H, Li T, Rayner S. MiRPara: a SVM-based software tool for prediction of most probable microRNA coding regions in genome scale sequences. BMC Bioinformatics. 2011;12(1):107. doi: 10.1186/1471-2105-12-107. [PMC free article] [PubMed] [Cross Ref]
Xu & Liang (2001) Xu Q-S, Liang Y-Z. Monte Carlo cross validation. Chemometrics and Intelligent Laboratory Systems. 2001;56(1):1–11. doi: 10.1016/S0169-7439(00)00122-2. [Cross Ref]
Xuan et al. (2011a) Xuan P, Guo M, Huang Y, Li W, Huang Y. MaturePred: efficient identification of microRNAs within novel plant pre-miRNAs. PLoS ONE. 2011a;6(11):e2135 doi: 10.1371/journal.pone.0027422. [PMC free article] [PubMed] [Cross Ref]
Xuan et al. (2011b) Xuan P, Guo M, Liu X, Huang Y, Huang Y. PlantMiRNAPred: efficient classification of real and pseudo plant pre-miRNAs. Bioinformatics. 2011b;27(10):1368–1376. doi: 10.1093/bioinformatics/btr153. [PubMed] [Cross Ref]
Xuan et al. (2011c) Xuan P, Guo MZ, Wang J, Wang CY, Liu XY, Liu Y. Genetic algorithm-based efficient feature selection for classification of pre-miRNAs. Genetics and Molecular Research. 2011c;10(2):588–603. doi: 10.4238/vol10-2gmr969. [PubMed] [Cross Ref]
Yousef, Allmer & Khalifa (2015) Yousef M, Allmer J, Khalifa W. Sequence motif-based one-class classifiers can achieve comparable accuracy to two-class learners for plant microRNA detection. Journal of Biomedical Science and Engineering. 2015;08(10):684–694. doi: 10.4236/jbise.2015.810065. [Cross Ref]
Yousef, Allmer & Khalifa (2016) Yousef M, Allmer J, Khalifa W. Proceedings of the 9th international joint conference on biomedical engineering systems and technologies. 2016. Feature selection for microRNA target prediction comparison of one-class feature selection methodologies; pp. 219–225.
Yousef, Allmer & Khalifa (2016) Yousef M, Allmer J, Khalifa W. Accurate plant microRNA prediction can be achieved using sequence motif features. Journal of Intelligent Learning Systems and Applications. 2016;8:9–22. doi: 10.4236/jilsa.2016.81002. [Cross Ref]
Yousef et al. (2008) Yousef M, Jung S, Showe LC, Showe MK. Learning from positive examples when the negative class is undetermined–microRNA gene identification. Algorithms for Molecular Biology. 2008;3:2. doi: 10.1186/1748-7188-3-2. [PMC free article] [PubMed] [Cross Ref]
Yousef et al. (2016) Yousef M, Saçar Demirci MD, Khalifa W, Allmer J. Feature selection has a large impact on one-class classification accuracy for MicroRNAs in plants. Advances in Bioinformatics. 2016;2016 Article 5670851. [PMC free article] [PubMed]
Zhang et al. (2010) Zhang Z, Yu J, Li D, Zhang Z, Liu F, Zhou X, Wang T, Ling Y, Su Z. PMRD: plant microRNA database. Nucleic Acids Research. 2010;38(Database issue):D806–D813. doi: 10.1093/nar/gkp818. [PMC free article] [PubMed] [Cross Ref]

Articles from PeerJ are provided here courtesy of PeerJ, Inc