FSVML AND GA-FSVML WRAPPER APPROACHES FOR GENE SELECTION AND CLASSIFICATION USING EXPRESSIONS OF VERY FEW GENES

REVATHY N.1*, BALASUBRAMANIAN R.2
1Department of Computer Applications, Karpagam College of Engineering, Coimbatore- 32, TN, India.
2PPG Institute of Technology, Coimbatore- 35, TN, India.
* Corresponding Author : revs_kn@rediffmail.com

Received : 15-03-2012     Accepted : 12-04-2012     Published : 30-04-2012
Volume : 4     Issue : 2       Pages : 85 - 91
Genetics 4.2 (2012):85-91
DOI : http://dx.doi.org/10.9735/0975-2862.4.2.85-91

Conflict of Interest : None declared

Cite - MLA : REVATHY N. and BALASUBRAMANIAN R. "FSVML AND GA-FSVML WRAPPER APPROACHES FOR GENE SELECTION AND CLASSIFICATION USING EXPRESSIONS OF VERY FEW GENES." International Journal of Genetics 4.2 (2012):85-91. http://dx.doi.org/10.9735/0975-2862.4.2.85-91

Cite - APA : REVATHY N., BALASUBRAMANIAN R. (2012). FSVML AND GA-FSVML WRAPPER APPROACHES FOR GENE SELECTION AND CLASSIFICATION USING EXPRESSIONS OF VERY FEW GENES. International Journal of Genetics, 4 (2), 85-91. http://dx.doi.org/10.9735/0975-2862.4.2.85-91

Cite - Chicago : REVATHY N. and BALASUBRAMANIAN R. "FSVML AND GA-FSVML WRAPPER APPROACHES FOR GENE SELECTION AND CLASSIFICATION USING EXPRESSIONS OF VERY FEW GENES." International Journal of Genetics 4, no. 2 (2012):85-91. http://dx.doi.org/10.9735/0975-2862.4.2.85-91

Copyright : © 2012, REVATHY N. and BALASUBRAMANIAN R., Published by Bioinfo Publications. This is an open-access article distributed under the terms of the Creative Commons Attribution License, which permits unrestricted use, distribution and reproduction in any medium, provided the original author and source are credited.

Abstract

Recently, Gene expression profiling by microarray technique has been effectively utilized for classification and diagnostic guessing of cancer nodules in the field of medical sciences. But the techniques used for cancer classification is still in its lower level. There are various drawbacks in the existing classification techniques such as low testing accuracy, high training time, unreliability, etc. Moreover, microarray data consists of a high degree of noise. Gene ranking techniques such as T-Score, ANOVA, etc are later proposed to overcome those problems. But those approaches will sometimes wrongly predict the rank when large database is used. To overcome these issues, this paper mainly focuses on the development of an effective feature selection and classification technique for microarray gene expression cancer diagnosis for provide significant accuracy, reliability and less error rate. In this paper, Wrapper feature selection approach called the GA-FSVML approach is used for the effective feature selection of genes. In FSVML, the RBF kernel function in SVM is trained using modified Levenberg Marquadt algorithm. This approach proposes a Fast SVM Learning (FSVML) technique for the classification tasks. The experiment is performed on lymphoma data set and the result shows the better accuracy of the proposed FSVML with GA-FSVML classification approach when compared to the standard existing approaches.

Keywords

Feature subset Selection, GA-FSVML, Support Vector Machine, Modified Levenberg Marquardt Learning, FSVML

Introduction

Microarray data analysis has been extensively employed in several analysis over a extensive range of biological domains which consists of cancer classification by class detection and prediction, recognition of the unknown effects of a particular therapy, cancer diagnosis, etc [1,2] . In this research, efficient neural network techniques are used with effective learning algorithms for providing significant cancer classification. The main goal of microarrays is hybridization between two DNA strands, the property of balancing nucleic acid series to chiefly pair with each other by figuring out hydrogen bonds between complementary nucleotide base pairs.Several machine learning techniques have been developed for the examination of microarray data [3,4] . The grouping of gene microarray method and machine learning technique assures new approaches into mechanisms of living schemes. An application field where these methods are likely to create key contributions is the identification of cancers depends on clinical phase and biological activities. Such classifications have a huge contribution on diagnosis and treatment.
Various recent investigations in the field of microarray have discussed the application of feature selection approaches to high-dimensional datasets. These feature selection approaches can be used to choose smaller subsets of interesting genes, supporting the analysis of statistical models while keeping the highest possible degree of the accuracy of models developed on the full dataset [5] . Occasionally, by facilitating statistical learning approaches to concentrate only on highly predictive genes while eliminating redundant variables and irrelevant noise, feature selection techniques can even enhance the accuracy of statistical models.
Feature selection techniques are often partitioned into two types namely filter and wrapper techniques. Filter techniques generally rank each gene individually by certain quality criterion (for instance, the p-value of t-test comparing two populations of interest with regard to the expression levels of the gene in the populations) and then select the subset of genes with the n highest quality criteria. Wrapper techniques employ a search algorithm to compute subsets of the variables as a group, rather than individually [6] . A thorough search through all subsets is clearly not possible—there could be around 225,000 variable subsets to consider. Therefore, these search approaches utilizes heuristics to direct their search towards promising candidates.
This approach uses the GA-FSVML based Wrapper approach. A feature subset selection is a process that can automatically selects a relevant subset of features and ignores the rest, thus resulting in a more comprehensive model. In particular, a Genetic Algorithm- Fast Support Vector Machine Learning (GA-FSVML) based “wrapper” approach for feature subset selection was applied to the gene data set. Then the feature selected genes are given to the classifier.
Recent investigations and explorations have described several new characteristic features and the practical applicability of Support Vector Machines (SVMs) in knowledge discovery and data mining. SVMs were widely used to discover informative patterns [7] .
Support Vector Machine (SVM) is one of the most extensively used machine learning approaches depending on the statistical learning theory, which uses the structural risk minimization inductive principle with the goal to obtain a good generalization from narrow dataset. A significant feature of the SVM is that this transformation need not be implemented to find out the separating hyperplane in the possibly very-high dimensional feature space, a kernel representation can be used for the purpose of determining the separating hyperplane, where the solution evaluated at the support vectors is written as a weighted sum of the values of certain kernel function.
Thus, in order to obtain significant overall results, this paper uses GA-FSVML feature selection approach for feature selection of the gene and the classifier used in this paper is Fast Support Vector Machine Learning approach [8,9] .

Related works

There are different techniques proposed by different authors for the prediction of cancer regions. Every technique has its own advantages and disadvantages. Some of the existing techniques are presented in this section.
Two vital problems in mammogram analysis for breast cancer in MR-images are described by Behnamghader et al., in [10] . The first is category is between normal and abnormal cases and then, classification between benign and malignant in cancerous cases. This proposed approach obtains textural and statistical illustrative features that are applied to a learning engine depending on the utilization of SVM learning framework to categorize them. The experimental observations provide significant accuracy in both classification problems, that shows the suitable interaction of the features and choosing powerful classifier i.e. SVM leads us to a brilliant outcome.
Jose et al., [11] suggested a Genetic Embedded Approach for gene selection and classification of microarray data classification which focuses on the selection of subsets of relevant genes to attain good classification performance. The author describes a genetic embedded technique that carry out choosing task for a SVM classifier. The key aspect of the proposed technique is that, it focuses the highly specialized crossover and mutation operators that consider the gene ranking information provided by the SVM classifier. The effectiveness of this approach is assessed using three well-known benchmark datasets from the literature, showing highly competitive results.
Rui et al., [12] proposed a multiclass cancer classification using semisupervised ellipsoid ARTMAP and particle swarm optimization with gene expression data [13] . It is critical for cancer prediction and treatment to perfectly categorize the site of origin of a cancer. With huge progress of DNA microarray techniques, creating gene expression profiles for various cancer kinds has previously turn out to be a capable way for cancer classification [14] . In addition to research on binary classification like normal versus tumor samples that focuses on various issues from a mixture of disciplines, the discrimination of multiple tumor kinds is also essential. In the meantime, the choosing of genes that are appropriate to definite cancer kinds not only enhances the performance of the classifiers, but also offers molecular insights for treatment. Here, the author utilizes the semisupervised ellipsoid ARTMAP (ssEAM) for multiclass cancer discrimination and particle swarm optimization for informative gene selection. ssEAM is a neural network technique [15] embedded in adaptive resonance theory and applicable for classification purpose. ssEAM characterizes fast, stable and finite learning and generates hyperellipsoidal clusters, containing complex nonlinear decision boundaries. PSO is an evolutionary algorithm-based method for global optimization. A discrete binary version of PSO is used to represent whether genes are selected or not. The effectiveness of ssEAM/PSO for multiclass cancer diagnosis is illustrated with the help of testing it on three publicly existing multiple-class cancer data sets.
Huilin et al., [16] presents the optimized kernel machines for cancer classification using gene expression data. This technique enhances the performances of the classifiers in classifying gene expression data. Intending to enhance the class separability of the data, the author uses a highly flexible kernel function model, the data-dependent kernel, as the objective kernel to be optimized.

Methodology

The important focus of this paper is to use the Fast Support Vector Learning algorithm in which the Radial Basis Function (Gaussian) kernel is trained using the modified Levenberg-Marquardt approach.

Fast SVM Learning (FSVML)

A novel Modified Levenberg-Marquardt like second-order algorithm for tuning the Parzen window σ in a Radial Basis Function (Gaussian) kernel is proposed in this paper. In this scenario, each property has its own sigma parameter connected with it. The values of the optimized σ are then utilized as a gauge for variable selection. Kernel Partial Least Squares (K-PLS) model is applied to a variety of benchmark data sets to assess the efficiency of the second-order sigma tuning process for an RBF kernel. The sigma-tuned RBF kernel model performs better than K-PLS and SVM models with a single sigma value.

Sigma Tuning Algorithm

Metric is selected as an error metric, represented as , which depends on the vector , Leave-One-Out (LOO) K-PLS is used to attain an initial , value based on an initial starting guess for the sigma-vector represented as , A second-order gradient descent technique is utilized to reduce the objective function , find the optimal choice for The search process initiates from the initial point , The value of , is updated depending on the minimization of the leave-one-out (or alternatively, leave several out) tuning (or validation) error, rather than directly diminishing the training error [Fig-1] .
According to Newton's rule for identifying a minimum in a multi-dimensional space, the relation between and at the minimum can be written as:

(1)

where H is the Hessian matrix. is a vertical vector, as expressed [Eq-2] (2)

After rearranging, the equation can be reorganized as

(3)

where . In order to efficiently proceed towards a converged solution, a Levenberg-Marquardt approach will be utilized.
Because of the approximate evaluation of the Hessian, a heuristic coefficient α will be introduced in the iterative updating procedure for the elements of σ leading to:

(4)

The value of α is set to 0.5 which turns out to be a robust choice based on hundreds of experiments with this algorithm on different datasets. Due to the drawbacks of Levenberg-Marquardt algorithm, this paper uses modified Levenberg Marquardt Algortihm.

Modified Levenberg Marquardt Algortihm

A Modified Levenberg-Marquardt algorithm is used for training the neural network [19] . Considering performance index is using the Newton method the equation obtained is as follows:

(5)

(6)

(7)

(8)

he gradient can write as:

(9)

Where

[Eq-10] (10)

is called the Jacobian matrix.

Then, the Hessian matrix is to be found. The k, j elements of the Hessian matrix yields as:

(11)

The Hessian matrix can then be expressed as follows:

(12)

(13)

If is small assumed, the Hessian matrix can be approximated as:

(14)

Using equations (4) and (12), the Gauss-Newton method is obtained as follows:



(15)

The advantage of Gauss-Newton is that it does not require calculation of second derivatives. There is a problem the Gauss-Newton method is the matrix may not be invertible. This can be overcome by using the following modification.
Hessian matrix can be written as:

(16)

Suppose that the eigen values and eigenvectors of H are and Then:







(17)

Therefore the eigenvectors of G are the same as the eigenvectors of H and the eigen values of G are . The matrix G is positive definite by increasing μ until for all i therefore the matrix will be invertible.
This leads to Levenberg-Marquardt algorithm:

(18)

(19)

As known, learning parameter, μ is illustrator of steps of actual output movement to desired output. In the standard LM method, μ is a constant number. This work modifies LM method using μ as:

(20)

Where e is a matrix therefore is a therefore is invertible.

Therefore, if actual output is far than desired output or similarly, errors are large so, it converges to desired output with large steps. Likewise, when measurement of error is small then, actual output approaches to desired output with soft steps. Thus, error oscillation reduces greatly. Modified Levenberg-Marquardt Algorithm for Learning is used for learning of the RBF kernel function in SVM which provides significant performance in cancer classification. This technique reduces the amount of time taken in learning procedure.

Microarray Gene Selection and Classification Approach

There are two phases included in the proposed technique. In the first phase, every gene in the training data are selected with the help a feature selection technique called GA-FSVML Wrapper feature selection approach in which the RBF kernel function in SVM is trained using Modified Levenberg Marquardt algorithm. In the second phase, the classification ability of every simple combination among the selected genes is tested with the help of a classifier called FSVML. Thus, both the selection and the classification technique provide better results.

Phase 1- GA-FSVML based Wrapper Feature Selection Approach

Feature subset selection is an optimization problem, which deals with searching the space of possible features to recognize one that is optimum or near-optimal with respect to certain performance measures (e.g., accuracy, learning time, etc.) Wrapper and filter feature selection approaches are available in literature. [Fig-2] shows the flowchart of GA-FSVML wrapper feature subset selection.
This approach uses the randomized wrapper feature selection approach. In particular, genetic algorithm paradigm is selected for randomization and FSVML as a base learner in wrapper approach. Alternatively, a population of feature subsets is developed via the process of genetic algorithm and a feature subset is computed via training and testing a FSVML with the data set. Genetic Algorithms (GAs) are stochastic search approaches based on the method of natural selection and genetics and are usually very effective for quick global search of large search spaces in complicated optimization issues. Earlier works have reported the feasibility of GA for wrapper approach to feature subset selection [17] . FSVML also suits as a base learner well because of its fast training ability. FSVML novelty detector was observed to produce equivalent performance with that of neural network; however, the learning time is much faster than that of neural network. An initial population (genes) is made up of diversified binary strings denoting the features selected. These genes undergo crossover and mutation, assessed by the FSVML base learner. Only those genes that are selected based on the particular multi-criteria fitness are put back into the population and the process is repeated for a fixed number of generations. The best solutions are obtained at the end of the complete iterations.
In the proposed GA-FSVML wrapper technique, a Gaussian kernel is utilized for the induction approach, i.e. SVM and the parameters were tuned via some heuristic technique. GA was employed with the following settings. The chromosome is a binary string where each bit represents whether the equivalent feature is available (1) or absent (0). The population size was usually set at 30, but when the population diversity resulted in an unacceptable performance, it was modified up to 50. The crossover rate of 0.6 and the mutation rate of 0.01-0.02 were adopted with equivalent methods being two-point crossover and uniform mutation, respectively. Selection offers the powerful force in the evolutionary process and the selection pressure is vital. At the primary stage of evolution, a low selection pressure is chosen for a wide exploration of the search space. At the end of evolution, where the population is near convergence, a high selection pressure is used to exploit the most promising regions of the search space. As for the sampling space, a regular one was selected which has the size of the particular population and is made up of all the offspring and only segments of parents. The sampling mechanism follows the probabilistic roulette wheel selection. To discriminate among the similar strong individuals in the last 10%-20% generations, a linear scaling technique was applied to handle the selection probability.
The fitness function integrated three different criteria, i.e. the accuracy of the novelty detector, the learning time used and the dimension reduction ratio. One definitions of the fitness function emphasized more on the accuracy:

(21)

Where denotes the fitness of the feature subset represented by x, Acc(x) represents the test accuracy of the FSVML novelty detector using the feature subset represented by x and is the time taken to train the FSVML.

Though, the test accuracy is the only vital criterion, dimension reduction ratio and training time is also included into the fitness function in that when the model show comparable results, the model with least training time which is vital in practical application and the feature subset with the smaller dimension which is less susceptible to introduce irrelevant or redundant features, are more preferred. In fact, proper tradeoff values among the multiple objectives have to be based on the knowledge of the problem domain or the experimental results.

Phase 2: Classification using Fast Support Vector Machines Learning (FSVML) Approach

SVM is a significant machine learning technique which is based on artificial intelligence. SVM is an efficient approach for training classifiers depending on various functions such as polynomial functions, radial basis functions, neural networks etc. In Support Vector Machine (SVM), the classifier is generated using a hyper linear separating plane.

To establish the boundary, two parallel hyperplanes are created, one on every side of the separating hyperplane between the two data sets. For SVM, a data point is represented as a p dimensional vector and it is required to distinguish whether it can split such points with a p−1-dimensional hyperplane. This is called a linear classifier.
To build a SVM classifier, a kernel function and its parameters need to be chosen. In this work, the following kernel function has been applied to build SVM classifiers:

Radial basis function



is the width of the function.
A kernel in Support Vector Machine is looked upon as a similarity measure to recode the input data. The kernel is used along with a map function. Identifying the best options for the kernel function and parameters is a challenging task, when applied to real dataset.
Usually, the recommended kernel function [18] for nonlinear problems is the Gaussian radial basis function, because it resembles the sigmoid kernel for certain parameters and it requires less parameters than a polynomial kernel. The kernel function parameter γ and the parameter C, which controls the complexity of the decision function versus the training error minimization, can be determined by running a 2 dimensional grid search, which means that the values for pairs of parameters (C, γ) are generated in a predefined interval with a fixed step. The performance of each combination is computed and used to determine the best pair of parameters.

Proposed Learning Algorithm for RBF Kernel

A novel Modified Levenberg-Marquardt like second-order algorithm for tuning the Parzen window σ in a Radial Basis Function (Gaussian) kernel is proposed in this paper. In this scenario, each property has its own sigma parameter connected with it. The values of the optimized σ are then utilized as a gauge for variable selection. Kernel Partial Least Squares (K-PLS) model is applied to a variety of benchmark data sets to assess the efficiency of the second-order sigma tuning process for an RBF kernel. The sigma-tuned RBF kernel model performs better than K-PLS and SVM models with a single sigma value.
Modified Levenberg-Marquardt Algorithm for Learning is used for learning of the RBF kernel function in SVM which provides significant performance in cancer classification which is described in section 3.1. This technique reduces the amount of time taken in learning procedure.

FSVML with GA-FSVML Algorithm Description

Step 1: The raw gene input data is given to the GA-FSVML feature selection technique.

Step 2: Features are selected based on the GA-FSVML based Wrapper Feature Selection Approach.

Step 3: Features selected genes are given as input to the SVM classifier.
Radial basis function is the width of the function.

Step 4: Modified Levenberg Marquat learning is used for training the RBF kernel in SVM.

As support vector machines are linear classifier that has the capability of finding the optimal hyper plane that increases the separation among patterns, this characteristic creates support vector machines as a potential means for gene expression data examination purposes. The 5 fold cross validation (CV) is performed for support vector machine in the training data set to adjust their constraints. First, the entire data set is split into training (F1) and testing (F2) data by random. The genes are ranked with the help of samples of F1. The combination (FC1) is produced with the help of 2 genes from 20. Then FC1 is arbitrarily split into 5 folds (fc1, fc2, fc3, fc4 and fc5). Among these folds one fold is chosen for testing. The other 4 folds are used as a classifier for FSVML. This combination is produces continuously and stops only when the better accuracy is achieved. At last with the fitted FSVML, the prediction can be carried out.

Experimental Results

The experimentation on the proposed method is carried on lymphoma data set. In the lymphoma data set, there are 42 samples obtained from Diffuse Large B-cell Lymphoma (DLBCL) [20] , nine samples from Follicular Lymphoma (FL) and 11 samples from Chronic Lymphocytic Leukemia (CLL). The whole dataset contains the expression data of 4026 genes. Some data may be lost in the dataset because of some error. For filling those lost values k-nearest neighbor technique is used.
Initially, the 62 samples are split randomly into 2 groups: 31 samples for testing, 31 samples for training. Based on the enrichment scores in the training set, the whole sets of 4026 genes are ranked. Then, 200 genes with highest rank are chosen. Finally, the genes are passed to the FSVML classifier for classification. [Fig-3] shows the resulted classification time for different gene samples. It can be observed that the proposed FSVML with GA-FSVML technique takes lesser time for classification when compared with other approaches like SVM with T-Score, SVM with GA-SVM and FSVML with GA-SVM approaches.
The comparison of accuracy of the classification approaches such as SVM with T-Score, SVM with GA-SVM, FSVML with GA-SVM and FSVML with GA-FSVML is shown in [Fig-4] . It is clear from the figure that the proposed FSVML with GA-FSVML technique resulted in better accuracy for all the samples used for classification.

Conclusion

The importance of cancer diagnosis and classification in the field of medical sciences has been increasing day by day. Several researches have been done in gene ranking and classifications to develop a novel approach as the existing approaches have lot of drawbacks such as low testing accuracy, high classification time, etc. This paper focuses on developing an efficient gene selection and classification techniques. The microarray gene data must be preprocessed for classification with significant accuracy using the classifier. The feature selection technique is used to support that task. This paper uses an efficient wrapper feature selection algorithm called GA-FSVML. The selected features from the GA-FSVML approach are given as input to the FSVML classifier in which the modified Levenberg-Marquardt Algorithm is used for learning of the RBF kernel function in SVM. Then the classifier is trained with that data. Finally, the classification of gene for identifying the cancer is performed. The experiment is performed with the help of lymphoma data set. The experimental result shows that the proposed FSVML with GA-FSVML technique results in better accuracy and consumes less time for classification when compared to other existing techniques taken into consideration.

References

[1] Tusher V.G., Tibshirani R. and Chu G. (2001) Significance Analysis of Microarrays applied to the Ionizing Radiation Response, 98 (51), 16-21.  
» CrossRef   » Google Scholar   » PubMed   » DOAJ   » CAS   » Scopus  

[2] Shalon T.D. (1995) DNA Microarrays: A New Tool for Genetic Analysis, Stanford University. Ph.D. thesis.  
» CrossRef   » Google Scholar   » PubMed   » DOAJ   » CAS   » Scopus  

[3] Brown M. (2000) The National Academy of Sciences, 97, 262-267.  
» CrossRef   » Google Scholar   » PubMed   » DOAJ   » CAS   » Scopus  

[4] Zhang H., Yu C., Singer B. and Xiong M. (2001) Recursive Partitioning for Tumor Classification with Gene Expression Microarray Data, 6730-6735.  
» CrossRef   » Google Scholar   » PubMed   » DOAJ   » CAS   » Scopus  

[5] Goh L., Song Q. and Kasabov N. (2004) The Second Asia-Pacific Conference on Bioinformatics, 161-166.  
» CrossRef   » Google Scholar   » PubMed   » DOAJ   » CAS   » Scopus  

[6] Enzhe Yu and Sungzoon Cho (2003) GA-SVM Wrapper Approach for Feature Subset Selection in Keystroke Dynamics Identity Verification.  
» CrossRef   » Google Scholar   » PubMed   » DOAJ   » CAS   » Scopus  

[7] Mallika R. and Saravanan V. (2010) World Academy Of Science, Engineering And Technology, 62.  
» CrossRef   » Google Scholar   » PubMed   » DOAJ   » CAS   » Scopus  

[8] Xiaogang Ruan, Jinlian Wang, Hui Li and Xiaoming Li. (2008)The 2nd International Conference on Bioinformatics and Biomedical Engineering, 342-346.  
» CrossRef   » Google Scholar   » PubMed   » DOAJ   » CAS   » Scopus  

[9] Furey T., Cristianini N., Duffy N., Bednarski D., Schummer M. and Haussler D. (2001) Support Vector Machine Classification and Validation of Cancer Tissue Samples using Microarray Expression Data.  
» CrossRef   » Google Scholar   » PubMed   » DOAJ   » CAS   » Scopus  

[10] Behnamghader E., Ardekani R.D., Fatemizadeh E. (2007) 5th International Symposium on Image and Signal Processing and Analysis (ISPA 2007), 98 - 101.  
» CrossRef   » Google Scholar   » PubMed   » DOAJ   » CAS   » Scopus  

[11] Hernandez J.C., Duval B. and Jin-Kao HaoGuyon, Weston J., Barnhill S. and Vapnik V. (2007) A Genetic Embedded Approach for Gene Selection and Classification of Microarray Data.  
» CrossRef   » Google Scholar   » PubMed   » DOAJ   » CAS   » Scopus  

[12] Berns A. (2000) Nature, 491-492.  
» CrossRef   » Google Scholar   » PubMed   » DOAJ   » CAS   » Scopus  

[13] Dubitzky W., Granzow M. and Berrar D. (2002) Comparing Symbolic and Subsymbolic Machine Learning Approaches to Classification of Cancer and Gene Indentification, Kluwer Academic.  
» CrossRef   » Google Scholar   » PubMed   » DOAJ   » CAS   » Scopus  

[14] Khan J., Wei J., Ringner M. and Saal L. (2001) Nature Medicine.  
» CrossRef   » Google Scholar   » PubMed   » DOAJ   » CAS   » Scopus  

[15] Huilin Xiong and Xue-Wen Chen (2005) The 2005 IEEE Symposium on Computational Intelligence in Bioinformatics and Computational Biology, 1-7.  
» CrossRef   » Google Scholar   » PubMed   » DOAJ   » CAS   » Scopus  

[16] Yang I. and Hanavar V. (1998) Feature subset selection wings a genetic algorithm.  
» CrossRef   » Google Scholar   » PubMed   » DOAJ   » CAS   » Scopus  

[17] Cristianini N. and Shawe -Taylor J. (2000) An Introduction to Support Vector Machines and Other Kernel based Learning Methods.  
» CrossRef   » Google Scholar   » PubMed   » DOAJ   » CAS   » Scopus  

[18] Suratgar A.A., Tavakoli M.B. and Hoseinabadi A. (2005) World Academy of Science, Engineering and Technology, 46-48.  
» CrossRef   » Google Scholar   » PubMed   » DOAJ   » CAS   » Scopus  

Images
Fig. 1- Process flow for Sigma Tuning
Fig. 2- GA-FSVML Wrapper Feature Subset Selection
Fig. 3- Classification Time for Different Gene Samples
Fig. 4- Accuracy of Classification for Different Gene Samples
Eq. 2-
Eq. 10-