A Survey on Feature Selection Methods in Microarray Gene Expression Data for Cancer Classification

 

Gunavathi C1*, Premalatha K2, Sivasubramanian K3

1School of Information Technology and Engineering, VIT University, Vellore, India.

2Department of CSE, Bannari Amman Institute of Technology, Sathyamangalam, India.

3Department of ECE, K.S. Rangasamy College of Technology, Tiruchengode, India.

*Corresponding Author E-mail: gunavathi.cm@vit.ac.in

 

ABSTRACT:

Microarray technology is commonly used in the study of disease diagnosis using gene expression levels. It not only received the attention of the research community but also has a wide range of applications. The success of microarray technology depends on the precision of measurement, the usage of tools in data mining, analytical methods and statistical modeling. The feature selection methods are used to find an informative representation, by removing noisy and irrelevant features which would improve the classification performance. There exist several works in the literature to select the significant features from the microarray. This paper reviews the feature selection methods used to select significant genes from the microarray gene expression data for cancer classification.

 

KEYWORDS:  Microarray, Feature Selection, Gene Expression, Cancer Classification, Gene Selection.

 

 

 

INTRODUCTION:

Microarray technology facilitates the assessment of Deoxyribonucleic acid (DNA) and Ribonucleic acid (RNA) variations. The raw microarray data are images that are transformed into gene expression matrices. The rows in the matrix correspond to genes, and the columns represent samples or trial conditions. The number in each cell signifies the expression level of a particular gene in a particular sample or condition1, 2. Expression levels can be absolute or relative. If two rows are similar, it implies that the respective genes are co-regulated and perhaps functionally related. By comparing samples, differentially expressed genes can be identified.

 

The main shortcoming of gene expression data is that it includes thousands of genes and a small number of samples. Ample methods and techniques have been proposed for cancer/tumour classification using microarray gene expression data. A number of gene selection methods have been introduced to select the informative genes for cancer/tumour prediction and diagnosis. Feature or gene selection methods can be used to mine the genes that directly involved in the classification and to eliminate irrelevant genes and improve the classification accuracy.

 

FEATURE SELECTION METHODS:

In this section the works related with gene selection and tumour classification using microarray gene expression data are discussed. Ding and Peng used minimum redundancy–maximum relevance framework to select the features from the microarray3.  This method is robust and generalizes unseen data and lead to significantly improved classifications in gene expression datasets. Two basic approaches for feature selection appear in machine learning and pattern recognition literature. A comparison between a group of different filter metrics and a wrapper sequential search procedure is carried out by Inza et al4. Duan et al proposed a feature selection method called Multiple SVM-recursive feature elimination (MSVM-RFE) that uses a backward elimination procedure5. This approach computes the feature ranking score from a statistical analysis of weight vectors of multiple linear SVMs trained on subsamples of the original training data at each step.

 

Yang et al proposed the methods for Gene selection (GS) namely GS1 and GS2 which are not affected by the unbalanced sample class sizes and did not assume any explicit statistical model on the gene expression values6. Self-organizing map (SOM) is used for clustering cancer data which is composed of important gene selection step7. This study presented a technique for cancer prediction from DNA microarray data. The prediction composed of two main steps which consist of the gene selection by statistic methodology and the clustering of cancer data using SOM.

 

An innovative generalization of Signal-to-noise ratio (SNR) for multiclass cancer discrimination through introduction of two indices, Gene dominant index and Gene dormant index (GDIs) is proposed by Tsai et al8. These two indices lead to the concepts of dominant and dormant genes with biological significance. The dominancy and dormancy of the identified biomarkers and their excellent discriminating power are demonstrated pictorially using the scatter plot of individual gene and 2-D Sammon's projection of the selected set of genes.

 

Liu et al proposed a gene selection method called Recursive feature addition (RFA) which combines supervised learning and statistical similarity measures9. Rough set concept with depended degrees is proposed by Wang and Gotoh10. In this method a small number of informative single gene and gene pairs are screened based on their depended degrees.

 

Chopra et al used gene pair combinations, called doublets, as input to the cancer classification algorithms, instead of the original expression values, and showed that the classification accuracy is consistently improved across different datasets and classification algorithms11. An Ensemble gene selection (EGS) method is proposed to choose multiple gene subsets for classification purpose, where the significant degree of gene is measured by conditional mutual information or its normalized form12. Different gene subsets have been obtained by setting different starting points of the search procedure. The subsets are used to train multiple base classifiers and aggregated into a consensus classifier by majority voting.

 

Yao and Li presented a method called Additive nonparametric margin maximum for case-based reasoning (ANMM4CBR) to obtain a robust classifier13. ANMM4CBR employed a Case-based reasoning (CBR) method for classification. In CBR, the rules that define the domain knowledge are difficult to obtain because only a small number of training samples are available. To select the most informative genes, feature selection via additively optimizing a nonparametric margin maximum criterion is proposed.

 

A feature selection approach based on statistically defined effective range of features for every class termed as Effective range based gene selection (ERGS) is proposed by Chandra and Gupta14. The basic principle behind ERGS is that higher weight is given to the feature that discriminates the classes clearly. Hyper-Box Enclosure (HBE) method based on mixed integer programming is used for the classification of some cancer types with a minimal set of predictor genes15.

 

Biomarker identifier (BMI) has been developed by Lee et al which identified the features with the ability to distinguish between two data groups of interest16. Margin Influence Analysis (MIA) is an approach designed to work with SVM for selecting informative genes17. The motivation for performing MIA lies in the fact that the margin of SVMs is an important factor which underlies the generalization performance of SVM models. The MIA could reveal genes which have statistically significant influence on the margin by using Mann-Whitney U test. Liu et al used a gene selection method called RFA, which combines supervised learning and statistical similarity measures18. To determine the final optimal gene set for prediction and classification, an algorithm named Lagging prediction peephole optimization (LPPO) is proposed.

 

A Group marker index (GMI), which is intuitive, of low-computational complexity gene selection method is presented by Tsai et al19. Most gene selection methods identify only single-class specific signature genes and cannot identify multiple-class specific signature genes easily. This method is efficient in the identification of both types of genes. This method is effective even when the sample size is small as well as when the class sizes are significantly different. Wang and Simon used a method based on a single gene to construct classification models20. The genes with the most powerful univariate class discrimination ability are identified and simple classification rules are constructed for class prediction using the single genes.

Alonso et al proposed a method that relaxed the maximum accuracy criterion to select the combination of attribute selection and classification algorithm21. Some suggestions are also given to choose a suitable combination of attribute selection and classifying algorithms for accuracy when using a low number of gene expressions. Huang et al presented an improved Semi-supervised local fisher discriminant (iSELF) analysis for gene expression data classification22. This work preserves the global structure of unlabeled samples in addition to separating labeled samples in different classes from each other.

 

Maji presented a quantitative measure based on mutual information that incorporates the information of sample categories to measure the similarity between attributes23. In this work, a supervised attribute clustering algorithm is used to find the groups of genes. It directly incorporates the information of sample categories into the attribute clustering process. The clusters are then refined incrementally based on sample categories. Sharma et al proposed an algorithm that divides genes into subsets first, the sizes of which are relatively small (roughly of size h), then selects informative smaller subsets of genes (of size r < h) from a subset and merges the chosen genes with another gene subset (of size r) to update the gene subset24. This process is repeated until all subsets are merged into one informative subset. Sharma et al proposed a null space based feature selection method for gene expression data in terms of supervised classification25. This method discards the redundant genes by applying the information of null space of scatter matrices.

 

Arevalillo and Navarro proposed a method named Maximum predictive–minimum redundancy (MPMR) for the selection of highly predictive genes having a low redundancy in their expression levels26. The predictive accuracy is assessed by Classification and regression trees (CART) models. Bhalla and Agrawal focused on the study that relaxes the maximum accuracy criterion for feature selection and selects the combination of feature selection method and classifier27. The method uses a small subset of features that gives the accuracy not statistically indicatively different than the maximum accuracy. By selecting the classifier that employs small number of features along with a good accuracy, the risk of over fitting (bias) is reduced.  Du et al proposed a feature selection method by considering all kinds of genes in the original gene set in a multi-step process28. The method eliminates the irrelevant, noisy and redundant genes and selects a subset of relevant genes at different stages.  Song et al proposed a fast clustering-based feature selection algorithm for gene selection from the microarray based on two steps29. In the first step, features are divided into clusters by using graph-theoretic clustering methods. In the second step, the most representative feature that is strongly related to target classes is selected from each cluster to form a subset of features. Xu et al proposed an improved SOM clustering algorithm which is based on neighborhood mutual information correlation measure30.

 

Genetic algorithm based feature selection with kNN is introduced by Gunavathi and Premalatha31. Feature selection prior to classification plays a vital role and a feature selection technique which combines Discrete wavelet transform (DWT) and Moving window technique (MWT) is used by Bennet et al32. In this method, kNN, NBC, and SVM are used as the classifiers. Maulik and Chakraborty proposed a Fuzzy preference based rough set (FPRS) method for feature (gene) selection with semi supervised SVMs33. The performance of the technique is compared with the SNR and Consistency based feature selection (CBFS) methods. Statistical methods and optimization algorithms are used for gene selection in Lung and Ovarian cancer by Gunavathi and Premalatha34.

 

Sharma et al proposed a feature selection method based on fixed-point algorithm for cancer classification using DNA microarray gene expression data35. In the fixed-point algorithm, between-class scatter matrix is used to compute the leading Eigen vector. This Eigen vector has been used to select genes. Sharma et al proposed a feature selection method using improved regularized linear discriminant analysis technique to select important genes36. Wang et al proposed an Online feature selection (OFS) method in which an online learner is permitted to retain a classifier involving only a small and fixed number of features37.

 

This work deals with two different tasks of online feature selection: 1) learning with full input, where a learner is allowed to access all the features to choose the subset of active features, and 2) learning with partial input, where only a limited number of features are permitted to be accessed for each instance by the learner. A comparative study on swarm intelligence techniques for feature selection in cancer classification is done in38. Feature selection based on cuckoo search optimization for cancer classification is used in39.

 

Table 1 shows the performance of various feature selection methods on different gene expression datasets used in the existing literature.

 

 

Table 1 Performance of the feature selection methods

S. No.

Reference

Methodology

Datasets used

Classification accuracy in percentage

1

Ding and Peng (2003)

Minimum redundancy–maximum relevance

Leukemia

Colon

NCI

Lung

Lymphoma

100.00

100.00

98.30

97.30

96.90

2

Inza et al (2004)

Comparison of filter and wrapper methods

Colon

Leukemia

95.16

100.00

3

Duan et al (2005)

MSVM-RFE

Breast

Colon

Leukemia

Lung

98.71

98.77

100.00

100.00

4

Yang et al (2006)

GS1 and GS2

Leukemia

SRBCT

GLIOMA

Lung

Carcinoma

MLL

Prostate

DLBCL

98.60

100.00

82.00

95.10

90.20

97.20

95.10

96.10

5

Vanichayobon et al (2007)

SOM

Carcinoma

Leukemia

Lung

100.00

100.00

100.00

6

Tsai et al (2008)

GDIs

SRBCT

Leukemia

CNS

Lung

97.80

94.70

78.40

96.55

7

Liu et al (2009)

RFA

MAQC-II breast cancer

MAQC-II Multiple Myeloma Dataset

90.6± 2.8

Not mentioned

8

Wang and Gotoh (2009)

Rough set

Leukemia

Lung

Prostate

Breast

Leukemia dataset 2

100.00

100.00

91.18

84.21

93.33

9

Chopra et al (2010)

Gene doublets

Colon

Leukemia

CNS

DLBCL

Lung

Prostate1

Prostate2

Prostate3

GCM

79.03

91.67

70.59

97.40

96.13

87.25

77.27

90.91

83.21

10

Liu et al (2010)

EGS

Breast

CNS

Leukemia

Lung

Lymphoma

Prostate

93.81

98.33

100.00

89.58

100.00

97.06

11

 

Yao and Li (2010)

ANMM4CBR

Leukemia

Colon

SRBCT

GCM

97.5 ± 1.7

86.7 ± 5.6

99.7 ± 0.3

63.3 ± 3.9

12

Chandra and Gupta (2011)

ERGS

ALL_AML

Colon

DLBCL

Lung

MLL

Prostate

100.00

83.87

97.92

100.00

97.22

94.12

13

Dagliyan et al (2011)

HBE

Leukemia

Prostate cancer

Prostate outcome

DLBCL

Lymphoma

SRBCT

100.00

96.08

95.24

96.10

97.87

100.00

14

Lee et al (2011)

BMI

MAQC-II

Airway dataset

92.98

82.99

 

Table 1 Continue…..

S. No.

Reference

Methodology

Datasets used

Classification accuracy in percentage

15

Li et al (2011)

MIA

Colon

Estrogen

100.00

100.00

16

Liu et al (2011)

RFA with LPPO

Leukemia

Lymphoma

Prostate

Colon

CNS

Breast

99.9

99.5

96.90

91.10

94.00

85.90

17

Tsai et al (2011)

GMI

SRBCT

Leukemia

CNS

Lung cancer

90.48

96.49

92.86

96.06

18

Wang and Simon (2011)

Single gene

Melanoma

Breast Cancer 1

Brain Cancer

Breast Cancer 2

Gastric Tumor

Lung Cancer 1

Lung Cancer 2

Lymphoma

Myeloma

Pancreatic Cancer

Prostate Cancer

97.00

69.00

80.00

58.00

89.00

98.00

93.00

74.00

68.00

90.00

89.00

19

Alonso et al (2012)

Combination of attribute selection and classification algorithm

ALL_AML

ALL

AML

Bladder

Brain

Breast

Breast-F

Breast-W

CNS

Colon

DLBCL

Leukemia

Lung

MLL

NCI60

Ovarian

Pancreatic

Prostate

100.00

94.18

77.20

96.67

55.56

70.96

100.00

92.56

75.49

88.41

98.67

68.33

99.33

98.21

84.34

100.00

92.50

93.08

20

Huang et al (2012)

iSELF

SRBCT

DLBCL

Brain Tumor

94.71 ± 0.46

91.33 ± 0.91

81.27 ± 0.72

21

Maji (2012)

Mutual information and supervised attribute clustering

Breast Cancer

Leukemia

Colon Cancer

Rheumatoid Arthritis versus Osteoarthritis

Rheumatoid Arthritis versus Healthy Controls

100.00

100.00

100.00

 

100.00

 

100.00

22

Sharma et al (2012a)

Gene subset

SRBCT

MLL

Prostate Tumor

100.00

100.00

100.00

23

Sharma et al (2012b)

Null space based feature selection

Acute Leukemia

SRBCT

MLL

Lung

100.00

100.00

100.00

97.3

24

Arevalillo and Navarro (2013)

MPMR

Colon

90.68

25

Bhalla and Agrawal (2013)

Combination of feature selection method and classifier

Breast

GCM

Lung

MLL

Lymphoma

Prostate

87.50

73.33

100.00

97.33

93.77

95.52

 

 

Table 1 Continue…….

S. No.

Reference

Methodology

Datasets used

Classification accuracy in percentage

26

Du et al (2013)

Multi-stage  method for feature selection

Leukemia

Prostate

Colon

Breast

Nervous

DLBCL

100.00

98.04

100.00

89.74

100.00

98.28

27

Song et al (2013)

Fast clustering-based feature selection algorithm

Colon

B-cell1

B-cell2

B-cell3

Leukemia1

Leukemia2

GCM

95.08

100.00

96.63

98.22

100.00

100.00

70.90

28

Xu et al (2013)

Improved SOM clustering algorithm

8D5K

AD400_10_10

Yeast

Heart

Class

Breast Cancer

78.10

76.50

67.90

73.70

71.10

72.90

29

Bennet et al (2014)

DWT and MWT

Colon

Ovarian

CNS

Leukemia

Breast

100.00

100.00

100.00

100.00

100.00


30

Maulik and  Chakraborty (2014)

FPRS with semi-supervised SVMs

Leukemia

SRBCT

MLL

DLBCL

Prostate

ALL

97.22 ± 3.21

95.61 ± 2.78

89.16 ± 3.32

95.25 ± 3.23

91.56 ± 4.39

85.27 ± 3.96

31

Sharma et al (2014a)

Fixed-point algorithm

SRBCT

MLL

ALL

70.00

100.00

94.00

32

Sharma et al (2014b)

Improved regularized linear discriminant analysis

SRBCT

MLL

Leukemia

100.00

100.00

100.00

33

Wang et al (2014)

OFS

Colon

99.92

 

 

 

CONCLUSION:

Feature selection is an important problem in large volumes of data for classification. The special nature of microarray data (large number of genes but small number of samples) creates the need for significant gene selection. There are numerous studies on feature selection for diagnosing cancer using microarray gene expression data. This paper has reviewed and analyzed the current research literature on the various methods for selecting the important features from the microarray for cancer classification. Extensive analysis on the comparison of classification accuracy of different datasets is also presented in this work.

 

CONFLICT OF INTEREST:

The authors declare no conflict of interest.

 

REFERENCES:

1.     Golub TR, Slonim DK, Tamayo P, Huard C, Gaasenbeek M, Mesirov JP, Coller H, Loh ML, Downing JR, Caligiuri MA, Bloomfield CD and Lander ES. Molecular classification of cancer: class discovery and class prediction by gene expression monitoring. Science. 1999; 286(5439): 531-537.

2.     Domany E. Cluster analysis of gene expression data. Journal of Statistical Physics. 2003; 110(3-6): 1117-1139.

3.     Ding C and Peng H. Minimum Redundancy Feature Selection from Microarray Gene Expression Data. Proceedings of the computational systems bioinformatics. 2003; 523-528.

4.     Inza I, Larranaga P, Blanco R and Cerrolaza AJ. Filter versus wrapper gene selection approaches in DNA microarray domains. Artificial Intelligence in Medicine. 2004; 31(2): 91-103.

5.     Duan KB, Rajapakse JC, Wang H and Azuaje F.  Multiple SVM-RFE for gene selection in cancer classification with expression data. IEEE Transactions on Nanobioscience. 2005; 4(3): 228-234.

6.     Yang K, Cai Z, Li J and Lin G. A stable gene selection in microarray data analysis. BMC Bioinformatics. 2006; 7: 228-243.

7.     Vanichayobon S, Siriphan W and Wiphada W. Microarray gene selection using self-organizing map. Proceedings of the seventh WSEAS international conference on simulation, modeling and optimization. 2007; 239-244.

8.     Tsai YS, Lin CT, Tseng GC, Chung IF and Pal NR. Discovery of dominant and dormant genes from expression data using a novel generalization of SNR for multi-class problems. BMC Bioinformatics. 2008; 9: doi: 10.1186/1471-2105/9/425.

9.     Liu Q, Sung AH, Chen Z, Liu J and Huang X. Feature selection and classification of MAQC-II breast cancer and multiple myeloma microarray gene expression data. PLoS ONE. 2009; 4(12): doi:10.1371/journal.pone.0008250.

10.  Wang X and Gotoh O. Accurate molecular classification of cancer using simple rules. BMC Medical genomics. 2009; 2: doi: 10.1186/1755-8794-2-64.

11.  Chopra P, Lee J, Kang J and Lee S. Improving cancer classification accuracy using gene pairs. PLoS ONE. 2010; 5(12): doi:10.1371/journal.pone.0014305.

12.  Liu H, Liu L and Zhang H. Ensemble gene selection for cancer classification. Pattern Recognition. 2010; 43(8): 2763-2772.

13.  Yao B and Li S. ANMM4CBR: a case-based reasoning method for gene expression data classification. Algorithms for Molecular Biology. 2010; 5(14): 1-11.

14.  Chandra B and Gupta M. An efficient statistical feature selection approach for classification of gene expression data. Journal of Biomedical Informatics. 2011; 44(4): 529-535.

15.  Dagliyan O, Uney YF, Kavakli IH and Turkay M. Optimization based tumor classification from microarray gene expression data. PLoS ONE. 2011; 6(2): doi:10.1371/journal.pone.0014579.

16.  Lee IH, Lushington GH and Visvanathan M. A filter-based feature selection approach for identifying potential biomarkers for lung cancer. Journal of Clinical Bioinformatics. 2011; 1(11): doi: 10.1186/2043-9113-1-11.

17.  Li HD, Liang YZ, Xu QS, Cao DS, Tan BB, Deng BC and Lin CC. Recipe for uncovering predictive genes using support vector machines based on model population analysis. IEEE/ACM Transactions on Computational Biology and Bioinformatics. 2011; 8(6): 1633-1641.

18.  Liu Q, Sung AH, Chen Z, Liu J, Chen L, Qiao M, Wang Z, Huang X and Deng Y. Gene selection and classification for cancer microarray data based on machine learning and similarity measures. BMC Genomics. 2011; 12 (1): 1-12.

19.  Tsai YS, Aguan K, Pal NR and Chung IF. Identification of single-and multiple-class specific signature genes from gene expression profiles by group marker index. PLoS ONE. 2011; 6(9): doi:10.1371/journal.pone.0024259.

20.  Wang X and Simon R. Microarray-based cancer prediction using single genes. BMC Bioinformatics. 2011; 12(391): doi: 10.1186/1471-2105-12-391.

21.  Alonso GCJ, Moro-Sancho IQ, Simon HA and Varela-Arrabal R. Microarray gene expression classification with few genes: criteria to combine attribute selection and classification methods. Expert Systems with Applications. 2012; 39(8): 7270-7280.

22.  Huang H, Li J and Liu J. Gene expression data classification based on improved semi-supervised local Fisher discriminate analysis. Expert Systems with Applications. 2012; 39(3): 2314-2320.

23.  Maji P. Mutual information-based supervised attribute clustering for microarray sample classification. IEEE Transactions on Knowledge and Data Engineering. 2012; 24(1): 127-140.

24.  Sharma A, Imoto S and Miyano S. A top-r feature selection algorithm for microarray gene expression data. IEEE/ACM Transactions on Computational Biology and Bioinformatics. 2012; 9(3): 754-764.

25.  Sharma A, Imoto S, Miyano S and Sharma V. Null space based feature selection method for gene expression data. International Journal of Machine Learning and Cybernetics. 2012; 3(4): 269-276.

26.  Arevalillo JM and Navarro H. Exploring correlations in gene expression microarray data for maximum predictive-minimum redundancy biomarker selection and classification. Computers in Biology and Medicine. 2013; 43(10): 1437-1443.

27.  Bhalla A and Agrawal RK. Microarray gene-expression data classification using less gene expressions by combining feature selection methods and classifiers. International Journal of Information Engineering and Electronic Business. 2013; 5: 42-48.

28.  Du W, Sun Y, Wang Y, Cao Z, Zhang C and Liang Y. A novel multi-stage feature selection method for microarray expression data analysis. International Journal of Data Mining and Bioinformatics; 2013; 7(1): 58-77.

29.  Song Q, Ni J and Wang G. A fast clustering-based feature subset selection algorithm for high-dimensional data. IEEE Transactions on Knowledge and Data Engineering. 2013; 25(1): 1-14.

30.  Xu J, Xu T, Sun L and Ren J. An improved correlation measure-based SOM clustering algorithm for gene selection. Journal of Software. 2013; 8(12): 3082-3087.

31.  Gunavathi C and Premalatha K. Performance analysis of genetic algorithm with kNN and SVM for feature selection in tumor classification. World Academy of Science, Engineering and Technology. International Journal of Computer, Information, Systems and Control Engineering. 2014; 8(8): 1357-1364.

32.  Bennet J, Ganaprakasam CA and Kannan A. A discrete wavelet based feature extraction and hybrid classification technique for microarray data analysis. The Scientific World Journal, 2014; Article ID 195470, doi:10.1155/2014/195470.

33.  Maulik U and Chakraborty D. Fuzzy preference based feature selection and semi supervised SVM for cancer classification. IEEE Transactions on Nan bioscience. 2014; 13(2): 152-160.

34.  Gunavathi C and Premalatha K. Statistical measures and optimization algorithms for gene selection in lung and ovarian tumor. World Academy of Science, Engineering and Technology. International Journal of Medical, Health, Pharmaceutical and Biomedical Engineering. 2014; 8(10): 685-691.

35.  Sharma A, Paliwal KK, Imoto S, Miyano S, Sharma V and Ananthanarayanan R. A Feature Selection Method using Fixed-Point Algorithm for DNA microarray gene expression data. International Journal of Knowledge-Based and Intelligent Engineering Systems. 2014; 18(1): 55-59.

36.  Sharma A, Paliwal KK, Imoto S and Miyano S. A feature selection method using improved regularized linear discriminate analysis. Machine Vision and Applications. 2014; 25(3): 775-786.

37.  Wang J, Zhao P, Hoi SCH and Jin R. Online feature selection and its applications. IEEE Transactions on Knowledge and Data Engineering. 2014; 26(3): 698-710.

38.  Gunavathi C and Premalatha K. A Comparative analysis of swarm intelligence techniques for feature selection in cancer classification. The Scientific World Journal. 2014; Article ID 693831, doi: 10.1155/2014/693831.

39.  Gunavathi C and Premalatha K. Cuckoo search optimization for feature selection in cancer classification: A new approach. International Journal of Data Mining and Bioinformatics. 2015; 13(3): 248-265.

 

 

 

 

 

Received on 06.03.2017             Modified on 02.04.2017

Accepted on 04.04.2017           © RJPT All right reserved

Research J. Pharm. and Tech. 2017; 10(5): 1395-1401.

DOI: 10.5958/0974-360X.2017.00249.9