A Survey on Feature Selection
Methods in Microarray Gene Expression Data for Cancer Classification
Gunavathi C1*,
Premalatha K2, Sivasubramanian K3
1School of
Information Technology and Engineering, VIT University, Vellore, India.
2Department of CSE,
Bannari Amman Institute of Technology, Sathyamangalam, India.
3Department of ECE,
K.S. Rangasamy College of Technology, Tiruchengode, India.
*Corresponding Author E-mail: gunavathi.cm@vit.ac.in
ABSTRACT:
Microarray
technology is commonly used in the study of disease diagnosis using gene
expression levels. It not only received the attention of the research
community but also has a wide range of applications. The success of microarray
technology depends on the precision of measurement, the usage of tools in data
mining, analytical methods and statistical modeling. The feature selection
methods are used to find an informative representation, by removing noisy and
irrelevant features which would improve the classification performance. There
exist several works in the literature to select the significant features from
the microarray. This paper reviews the feature selection methods used to select
significant genes from the microarray gene expression data for cancer
classification.
KEYWORDS: Microarray, Feature
Selection, Gene Expression, Cancer Classification, Gene Selection.
INTRODUCTION:
Microarray technology facilitates the
assessment of Deoxyribonucleic acid (DNA) and Ribonucleic acid (RNA)
variations. The raw microarray data are images that are transformed into gene
expression matrices. The rows in the matrix correspond to genes, and the
columns represent samples or trial conditions. The number in each cell
signifies the expression level of a particular gene in a particular sample or
condition1, 2. Expression levels can be absolute or relative. If two
rows are similar, it implies that the respective genes are co-regulated and
perhaps functionally related. By comparing samples, differentially expressed genes
can be identified.
The main shortcoming of gene
expression data is that it includes thousands of genes and a small number of
samples. Ample methods and techniques have been proposed for cancer/tumour
classification using microarray gene expression data. A number of gene
selection methods have been introduced to select the informative genes for
cancer/tumour prediction and diagnosis. Feature or gene selection methods can
be used to mine the genes that directly involved in the classification and to eliminate
irrelevant genes and improve the classification accuracy.
FEATURE SELECTION
METHODS:
In this section the works
related with gene selection and tumour classification using microarray gene
expression data are discussed. Ding and Peng used minimum redundancy–maximum
relevance framework to select the features from the microarray3.
This method is robust and generalizes unseen data and lead to significantly
improved classifications in gene expression datasets. Two basic approaches for
feature selection appear in machine learning and pattern recognition
literature. A comparison between a group of different filter metrics and a
wrapper sequential search procedure is carried out by Inza et al4.
Duan et al proposed a feature selection method called Multiple SVM-recursive
feature elimination (MSVM-RFE) that uses a backward elimination procedure5.
This approach computes the feature ranking score from a statistical analysis of
weight vectors of multiple linear SVMs trained on subsamples of the original
training data at each step.
Yang et al proposed the
methods for Gene selection (GS) namely GS1 and GS2 which are not affected by
the unbalanced sample class sizes and did not assume any explicit statistical
model on the gene expression values6. Self-organizing map (SOM) is
used for clustering cancer data which is composed of important gene selection
step7. This study presented a technique for cancer prediction from
DNA microarray data. The prediction composed of two main steps which consist of
the gene selection by statistic methodology and the clustering of cancer data
using SOM.
An innovative generalization
of Signal-to-noise ratio (SNR) for multiclass cancer discrimination through
introduction of two indices, Gene dominant index and Gene dormant index (GDIs) is
proposed by Tsai et al8. These two indices lead to the concepts of
dominant and dormant genes with biological significance. The dominancy and
dormancy of the identified biomarkers and their excellent discriminating power
are demonstrated pictorially using the scatter plot of individual gene and 2-D
Sammon's projection of the selected set of genes.
Liu et al proposed a gene
selection method called Recursive feature addition (RFA) which combines
supervised learning and statistical similarity measures9. Rough set
concept with depended degrees is proposed by Wang and Gotoh10. In
this method a small number of informative single gene and gene pairs are
screened based on their depended degrees.
Chopra et al used gene pair
combinations, called doublets, as input to the cancer classification
algorithms, instead of the original expression values, and showed that the
classification accuracy is consistently improved across different datasets and
classification algorithms11. An Ensemble gene selection (EGS) method
is proposed to choose multiple gene subsets for classification purpose, where
the significant degree of gene is measured by conditional mutual information or
its normalized form12. Different gene subsets have been obtained by
setting different starting points of the search procedure. The subsets are used
to train multiple base classifiers and aggregated into a consensus classifier
by majority voting.
Yao and Li presented a method
called Additive nonparametric margin maximum for case-based reasoning (ANMM4CBR)
to obtain a robust classifier13. ANMM4CBR employed a Case-based
reasoning (CBR) method for classification. In CBR, the rules that define the
domain knowledge are difficult to obtain because only a small number of
training samples are available. To select the most informative genes, feature
selection via additively optimizing a nonparametric margin maximum criterion is
proposed.
A feature selection approach
based on statistically defined effective range of features for every class
termed as Effective range based gene selection (ERGS) is proposed by Chandra
and Gupta14. The basic principle behind ERGS is that higher weight
is given to the feature that discriminates the classes clearly. Hyper-Box
Enclosure (HBE) method based on mixed integer programming is used for the
classification of some cancer types with a minimal set of predictor genes15.
Biomarker identifier (BMI) has
been developed by Lee et al which identified the features with the ability to
distinguish between two data groups of interest16. Margin Influence
Analysis (MIA) is an approach designed to work with SVM for selecting
informative genes17. The motivation for performing MIA lies in the
fact that the margin of SVMs is an important factor which underlies the
generalization performance of SVM models. The MIA could reveal genes which have
statistically significant influence on the margin by using Mann-Whitney U test.
Liu et al used a gene selection method called RFA, which combines supervised learning
and statistical similarity measures18. To determine the final
optimal gene set for prediction and classification, an algorithm named Lagging
prediction peephole optimization (LPPO) is proposed.
A Group marker index (GMI),
which is intuitive, of low-computational complexity gene selection method is
presented by Tsai et al19. Most gene selection methods identify only
single-class specific signature genes and cannot identify multiple-class
specific signature genes easily. This method is efficient in the identification
of both types of genes. This method is effective even when the sample size is
small as well as when the class sizes are significantly different. Wang and
Simon used a method based on a single gene to construct classification models20.
The genes with the most powerful univariate class discrimination ability are
identified and simple classification rules are constructed for class prediction
using the single genes.
Alonso et al proposed a method
that relaxed the maximum accuracy criterion to select the combination of
attribute selection and classification algorithm21. Some suggestions
are also given to choose a suitable combination of attribute selection and
classifying algorithms for accuracy when using a low number of gene
expressions. Huang et al presented an improved Semi-supervised local fisher
discriminant (iSELF) analysis for gene expression data classification22.
This work preserves the global structure of unlabeled samples in addition to
separating labeled samples in different classes from each other.
Maji presented a quantitative
measure based on mutual information that incorporates the information of sample
categories to measure the similarity between attributes23. In this
work, a supervised attribute clustering algorithm is used to find the groups of
genes. It directly incorporates the information of sample categories into the
attribute clustering process. The clusters are then refined incrementally based
on sample categories. Sharma et al proposed an algorithm that divides genes into
subsets first, the sizes of which are relatively small (roughly of size h),
then selects informative smaller subsets of genes (of size r < h) from a
subset and merges the chosen genes with another gene subset (of size r) to
update the gene subset24. This process is repeated until all subsets
are merged into one informative subset. Sharma et al proposed a null space
based feature selection method for gene expression data in terms of supervised
classification25. This method discards the redundant genes by applying
the information of null space of scatter matrices.
Arevalillo and Navarro
proposed a method named Maximum predictive–minimum redundancy (MPMR) for the
selection of highly predictive genes having a low redundancy in their
expression levels26. The predictive accuracy is assessed by
Classification and regression trees (CART) models. Bhalla and Agrawal focused
on the study that relaxes the maximum accuracy criterion for feature selection
and selects the combination of feature selection method and classifier27.
The method uses a small subset of features that gives the accuracy not
statistically indicatively different than the maximum accuracy. By selecting
the classifier that employs small number of features along with a good
accuracy, the risk of over fitting (bias) is reduced. Du et al proposed a
feature selection method by considering all kinds of genes in the original gene
set in a multi-step process28. The method eliminates the irrelevant,
noisy and redundant genes and selects a subset of relevant genes at different
stages. Song et al proposed a fast clustering-based feature selection
algorithm for gene selection from the microarray based on two steps29.
In the first step, features are divided into clusters by using graph-theoretic
clustering methods. In the second step, the most representative feature that is
strongly related to target classes is selected from each cluster to form a
subset of features. Xu et al proposed an improved SOM clustering algorithm
which is based on neighborhood mutual information correlation measure30.
Genetic algorithm based
feature selection with kNN is introduced by Gunavathi and Premalatha31.
Feature selection prior to classification plays a vital role and a feature
selection technique which combines Discrete wavelet transform (DWT) and Moving
window technique (MWT) is used by Bennet et al32. In this method,
kNN, NBC, and SVM are used as the classifiers. Maulik and Chakraborty proposed
a Fuzzy preference based rough set (FPRS) method for feature (gene) selection
with semi supervised SVMs33. The performance of the technique is
compared with the SNR and Consistency based feature selection (CBFS) methods.
Statistical methods and optimization algorithms are used for gene selection in
Lung and Ovarian cancer by Gunavathi and Premalatha34.
Sharma et al proposed a
feature selection method based on fixed-point algorithm for cancer
classification using DNA microarray gene expression data35. In the
fixed-point algorithm, between-class scatter matrix is used to compute the leading
Eigen vector. This Eigen vector has been used to select genes. Sharma et al
proposed a feature selection method using improved regularized linear
discriminant analysis technique to select important genes36. Wang et
al proposed an Online feature selection (OFS) method in which an online learner
is permitted to retain a classifier involving only a small and fixed number of
features37.
This work deals with two
different tasks of online feature selection: 1) learning with full input, where
a learner is allowed to access all the features to choose the subset of active
features, and 2) learning with partial input, where only a limited number of
features are permitted to be accessed for each instance by the learner. A
comparative study on swarm intelligence techniques for feature selection in
cancer classification is done in38. Feature selection based on
cuckoo search optimization for cancer classification is used in39.
Table 1 shows the performance
of various feature selection methods on different gene expression datasets used
in the existing literature.
Table 1 Performance of the
feature selection methods
|
S. No. |
Reference |
Methodology |
Datasets used |
Classification accuracy in percentage |
|
1 |
Ding and Peng (2003) |
Minimum redundancy–maximum relevance |
Leukemia Colon NCI Lung Lymphoma |
100.00 100.00 98.30 97.30 96.90 |
|
2 |
Inza et al (2004) |
Comparison of filter and wrapper methods |
Colon Leukemia |
95.16 100.00 |
|
3 |
Duan et al (2005) |
MSVM-RFE |
Breast Colon Leukemia Lung |
98.71 98.77 100.00 100.00 |
|
4 |
Yang et al (2006) |
GS1 and GS2 |
Leukemia SRBCT GLIOMA Carcinoma MLL Prostate DLBCL |
98.60 100.00 82.00 95.10 90.20 97.20 95.10 96.10 |
|
5 |
Vanichayobon et al (2007) |
SOM |
Carcinoma Leukemia Lung |
100.00 100.00 100.00 |
|
6 |
Tsai et al (2008) |
GDIs |
SRBCT Leukemia CNS Lung |
97.80 94.70 78.40 96.55 |
|
7 |
Liu et al (2009) |
RFA |
MAQC-II breast cancer MAQC-II Multiple Myeloma Dataset |
90.6± 2.8 Not mentioned |
|
8 |
Wang and Gotoh (2009) |
Rough set |
Leukemia Lung Prostate Breast Leukemia dataset 2 |
100.00 100.00 91.18 84.21 93.33 |
|
9 |
Chopra et al (2010) |
Gene doublets |
Colon Leukemia CNS DLBCL Lung Prostate1 Prostate2 Prostate3 GCM |
79.03 91.67 70.59 97.40 96.13 87.25 77.27 90.91 83.21 |
|
10 |
Liu et al (2010) |
EGS |
Breast CNS Leukemia Lung Lymphoma Prostate |
93.81 98.33 100.00 89.58 100.00 97.06 |
|
11
|
Yao and Li (2010) |
ANMM4CBR |
Leukemia Colon SRBCT GCM |
97.5 ± 1.7 86.7 ± 5.6 99.7 ± 0.3 63.3 ± 3.9 |
|
12 |
Chandra and Gupta (2011) |
ERGS |
ALL_AML Colon DLBCL Lung MLL Prostate |
100.00 83.87 97.92 100.00 97.22 94.12 |
|
13 |
Dagliyan et al (2011) |
HBE |
Leukemia Prostate cancer Prostate outcome DLBCL Lymphoma SRBCT |
100.00 96.08 95.24 96.10 97.87 100.00 |
|
14 |
Lee et al (2011) |
BMI |
MAQC-II Airway dataset |
92.98 82.99 |
Table 1 Continue…..
|
S. No. |
Reference |
Methodology |
Datasets used |
Classification accuracy in percentage |
|
15 |
Li et al (2011) |
MIA |
Colon Estrogen |
100.00 100.00 |
|
16 |
Liu et al (2011) |
RFA with LPPO |
Leukemia Lymphoma Prostate Colon CNS Breast |
99.9 99.5 96.90 91.10 94.00 85.90 |
|
17 |
Tsai et al (2011) |
GMI |
SRBCT Leukemia CNS Lung cancer |
90.48 96.49 92.86 96.06 |
|
18 |
Wang and Simon (2011) |
Single gene |
Melanoma Breast Cancer 1 Brain Cancer Breast Cancer 2 Gastric Tumor Lung Cancer 1 Lung Cancer 2 Lymphoma Myeloma Pancreatic Cancer Prostate Cancer |
97.00 69.00 80.00 58.00 89.00 98.00 93.00 74.00 68.00 90.00 89.00 |
|
19 |
Alonso et al (2012) |
Combination of attribute selection and classification algorithm |
ALL_AML ALL AML Bladder Brain Breast Breast-F Breast-W CNS Colon DLBCL Leukemia Lung MLL NCI60 Ovarian Pancreatic Prostate |
100.00 94.18 77.20 96.67 55.56 70.96 100.00 92.56 75.49 88.41 98.67 68.33 99.33 98.21 84.34 100.00 92.50 93.08 |
|
20 |
Huang et al (2012) |
iSELF |
SRBCT DLBCL Brain Tumor |
94.71 ± 0.46 91.33 ± 0.91 81.27 ± 0.72 |
|
21 |
Maji (2012) |
Mutual information and supervised attribute clustering |
Breast Cancer Leukemia Colon Cancer Rheumatoid Arthritis versus Osteoarthritis Rheumatoid Arthritis versus Healthy Controls |
100.00 100.00 100.00
100.00
100.00 |
|
22 |
Sharma et al (2012a) |
Gene subset |
SRBCT MLL Prostate Tumor |
100.00 100.00 100.00 |
|
23 |
Sharma et al (2012b) |
Null space based feature selection |
Acute Leukemia SRBCT MLL Lung |
100.00 100.00 100.00 97.3 |
|
24 |
Arevalillo and Navarro (2013) |
MPMR |
Colon |
90.68 |
|
25 |
Bhalla and Agrawal (2013) |
Combination of feature selection method and classifier |
Breast GCM Lung MLL Lymphoma Prostate |
87.50 73.33 100.00 97.33 93.77 95.52 |
Table 1 Continue…….
|
S. No. |
Reference |
Methodology |
Datasets used |
Classification accuracy in percentage |
|
26 |
Du et al (2013) |
Multi-stage method for feature selection |
Leukemia Prostate Colon Breast Nervous DLBCL |
100.00 98.04 100.00 89.74 100.00 98.28 |
|
27 |
Song et al (2013) |
Fast clustering-based feature selection algorithm |
Colon B-cell1 B-cell2 B-cell3 Leukemia1 Leukemia2 GCM |
95.08 100.00 96.63 98.22 100.00 100.00 70.90 |
|
28 |
Xu et al (2013) |
Improved SOM clustering algorithm |
8D5K AD400_10_10 Yeast Heart Class Breast Cancer |
78.10 76.50 67.90 73.70 71.10 72.90 |
|
29 |
Bennet et al (2014) |
DWT and MWT |
Colon Ovarian CNS Leukemia Breast |
100.00 100.00 100.00 100.00 100.00 |
30 |
Maulik and Chakraborty (2014) |
FPRS with semi-supervised SVMs |
Leukemia SRBCT MLL DLBCL Prostate ALL |
97.22 ± 3.21 95.61 ± 2.78 89.16 ± 3.32 95.25 ± 3.23 91.56 ± 4.39 85.27 ± 3.96 |
|
31 |
Sharma et al (2014a) |
Fixed-point algorithm |
SRBCT MLL ALL |
70.00 100.00 94.00 |
|
32 |
Sharma et al (2014b) |
Improved regularized linear discriminant analysis |
SRBCT MLL Leukemia |
100.00 100.00 100.00 |
|
33 |
Wang et al (2014) |
OFS |
Colon |
99.92 |
CONCLUSION:
Feature selection is an
important problem in large volumes of data for classification. The special
nature of microarray data (large number of genes but small number of samples)
creates the need for significant gene selection. There are numerous studies on
feature selection for diagnosing cancer using microarray gene expression data.
This paper has reviewed and analyzed the current research literature on the
various methods for selecting the important features from the microarray for
cancer classification. Extensive analysis on the comparison of classification
accuracy of different datasets is also presented in this work.
CONFLICT OF INTEREST:
The authors declare no conflict of interest.
REFERENCES:
1.
Golub
TR, Slonim DK, Tamayo P, Huard C, Gaasenbeek M, Mesirov JP, Coller H, Loh ML,
Downing JR, Caligiuri MA, Bloomfield CD and Lander ES. Molecular classification
of cancer: class discovery and class prediction by gene expression monitoring.
Science. 1999; 286(5439): 531-537.
2.
Domany
E. Cluster analysis of gene expression data. Journal of Statistical Physics.
2003; 110(3-6): 1117-1139.
3.
Ding C
and Peng H. Minimum Redundancy Feature Selection from Microarray Gene
Expression Data. Proceedings of the computational systems bioinformatics. 2003;
523-528.
4.
Inza I,
Larranaga P, Blanco R and Cerrolaza AJ. Filter versus wrapper gene selection
approaches in DNA microarray domains. Artificial Intelligence in Medicine.
2004; 31(2): 91-103.
5.
Duan
KB, Rajapakse JC, Wang H and Azuaje F. Multiple SVM-RFE for gene
selection in cancer classification with expression data. IEEE Transactions on
Nanobioscience. 2005; 4(3): 228-234.
6.
Yang K,
Cai Z, Li J and Lin G. A stable gene selection in microarray data analysis. BMC
Bioinformatics. 2006; 7: 228-243.
7.
Vanichayobon
S, Siriphan W and Wiphada W. Microarray gene selection using self-organizing
map. Proceedings of the seventh WSEAS international conference on simulation,
modeling and optimization. 2007; 239-244.
8.
Tsai
YS, Lin CT, Tseng GC, Chung IF and Pal NR. Discovery of dominant and dormant
genes from expression data using a novel generalization of SNR for multi-class
problems. BMC Bioinformatics. 2008; 9: doi: 10.1186/1471-2105/9/425.
9.
Liu Q,
Sung AH, Chen Z, Liu J and Huang X. Feature selection and classification of
MAQC-II breast cancer and multiple myeloma microarray gene expression data.
PLoS ONE. 2009; 4(12): doi:10.1371/journal.pone.0008250.
10. Wang X and Gotoh O.
Accurate molecular classification of cancer using simple rules. BMC Medical
genomics. 2009; 2: doi: 10.1186/1755-8794-2-64.
11. Chopra P, Lee J,
Kang J and Lee S. Improving cancer classification accuracy using gene pairs.
PLoS ONE. 2010; 5(12): doi:10.1371/journal.pone.0014305.
12. Liu H, Liu L and
Zhang H. Ensemble gene selection for cancer classification. Pattern
Recognition. 2010; 43(8): 2763-2772.
13. Yao B and Li S.
ANMM4CBR: a case-based reasoning method for gene expression data
classification. Algorithms for Molecular Biology. 2010; 5(14): 1-11.
14. Chandra B and Gupta
M. An efficient statistical feature selection approach for classification of
gene expression data. Journal of Biomedical Informatics. 2011; 44(4): 529-535.
15. Dagliyan O, Uney YF,
Kavakli IH and Turkay M. Optimization based tumor classification from
microarray gene expression data. PLoS ONE. 2011; 6(2):
doi:10.1371/journal.pone.0014579.
16. Lee IH, Lushington
GH and Visvanathan M. A filter-based feature selection approach for identifying
potential biomarkers for lung cancer. Journal of Clinical Bioinformatics. 2011;
1(11): doi: 10.1186/2043-9113-1-11.
17. Li HD, Liang YZ, Xu
QS, Cao DS, Tan BB, Deng BC and Lin CC. Recipe for uncovering predictive genes
using support vector machines based on model population analysis. IEEE/ACM
Transactions on Computational Biology and Bioinformatics. 2011; 8(6):
1633-1641.
18. Liu Q, Sung AH, Chen
Z, Liu J, Chen L, Qiao M, Wang Z, Huang X and Deng Y. Gene selection and
classification for cancer microarray data based on machine learning and
similarity measures. BMC Genomics. 2011; 12 (1): 1-12.
19. Tsai YS, Aguan K,
Pal NR and Chung IF. Identification of single-and multiple-class specific
signature genes from gene expression profiles by group marker index. PLoS ONE.
2011; 6(9): doi:10.1371/journal.pone.0024259.
20. Wang X and Simon R.
Microarray-based cancer prediction using single genes. BMC Bioinformatics.
2011; 12(391): doi: 10.1186/1471-2105-12-391.
21. Alonso GCJ,
Moro-Sancho IQ, Simon HA and Varela-Arrabal R. Microarray gene expression
classification with few genes: criteria to combine attribute selection and
classification methods. Expert Systems with Applications. 2012; 39(8):
7270-7280.
22. Huang H, Li J and
Liu J. Gene expression data classification based on improved semi-supervised
local Fisher discriminate analysis. Expert Systems with Applications. 2012;
39(3): 2314-2320.
23. Maji P. Mutual
information-based supervised attribute clustering for microarray sample
classification. IEEE Transactions on Knowledge and Data Engineering. 2012;
24(1): 127-140.
24. Sharma A, Imoto S
and Miyano S. A top-r feature selection algorithm for microarray gene
expression data. IEEE/ACM Transactions on Computational Biology and
Bioinformatics. 2012; 9(3): 754-764.
25. Sharma A, Imoto S,
Miyano S and Sharma V. Null space based feature selection method for gene
expression data. International Journal of Machine Learning and Cybernetics.
2012; 3(4): 269-276.
26. Arevalillo JM and
Navarro H. Exploring correlations in gene expression microarray data for
maximum predictive-minimum redundancy biomarker selection and classification.
Computers in Biology and Medicine. 2013; 43(10): 1437-1443.
27. Bhalla A and Agrawal
RK. Microarray gene-expression data classification using less gene expressions
by combining feature selection methods and classifiers. International Journal
of Information Engineering and Electronic Business. 2013; 5: 42-48.
28. Du W, Sun Y, Wang Y,
Cao Z, Zhang C and Liang Y. A novel multi-stage feature selection method for
microarray expression data analysis. International Journal of Data Mining and
Bioinformatics; 2013; 7(1): 58-77.
29. Song Q, Ni J and
Wang G. A fast clustering-based feature subset selection algorithm for
high-dimensional data. IEEE Transactions on Knowledge and Data Engineering.
2013; 25(1): 1-14.
30. Xu J, Xu T, Sun L
and Ren J. An improved correlation measure-based SOM clustering algorithm for
gene selection. Journal of Software. 2013; 8(12): 3082-3087.
31. Gunavathi C and
Premalatha K. Performance analysis of genetic algorithm with kNN and SVM for
feature selection in tumor classification. World Academy of Science,
Engineering and Technology. International Journal of Computer, Information,
Systems and Control Engineering. 2014; 8(8): 1357-1364.
32. Bennet J,
Ganaprakasam CA and Kannan A. A discrete wavelet based feature extraction and
hybrid classification technique for microarray data analysis. The Scientific
World Journal, 2014; Article ID 195470, doi:10.1155/2014/195470.
33. Maulik U and
Chakraborty D. Fuzzy preference based feature selection and semi supervised SVM
for cancer classification. IEEE Transactions on Nan bioscience. 2014; 13(2):
152-160.
34. Gunavathi C and
Premalatha K. Statistical measures and optimization algorithms for gene
selection in lung and ovarian tumor. World Academy of Science, Engineering and
Technology. International Journal of Medical, Health, Pharmaceutical and
Biomedical Engineering. 2014; 8(10): 685-691.
35. Sharma A, Paliwal
KK, Imoto S, Miyano S, Sharma V and Ananthanarayanan R. A Feature Selection
Method using Fixed-Point Algorithm for DNA microarray gene expression data.
International Journal of Knowledge-Based and Intelligent Engineering Systems.
2014; 18(1): 55-59.
36. Sharma A, Paliwal
KK, Imoto S and Miyano S. A feature selection method using improved regularized
linear discriminate analysis. Machine Vision and Applications. 2014; 25(3):
775-786.
37. Wang J, Zhao P, Hoi
SCH and Jin R. Online feature selection and its applications. IEEE Transactions
on Knowledge and Data Engineering. 2014; 26(3): 698-710.
38. Gunavathi C and
Premalatha K. A Comparative analysis of swarm intelligence techniques for
feature selection in cancer classification. The Scientific World Journal. 2014;
Article ID 693831, doi: 10.1155/2014/693831.
39. Gunavathi C and
Premalatha K. Cuckoo search optimization for feature selection in cancer
classification: A new approach. International Journal of Data Mining and
Bioinformatics. 2015; 13(3): 248-265.
Received on
06.03.2017
Modified on 02.04.2017
Accepted on
04.04.2017 © RJPT All
right reserved
Research J. Pharm. and Tech.
2017; 10(5): 1395-1401.
DOI: 10.5958/0974-360X.2017.00249.9