Skip to main content

A graph-based gene selection method for medical diagnosis problems using a many-objective PSO algorithm

Abstract

Background

Gene expression data play an important role in bioinformatics applications. Although there may be a large number of features in such data, they mainly tend to contain only a few samples. This can negatively impact the performance of data mining and machine learning algorithms. One of the most effective approaches to alleviate this problem is to use gene selection methods. The aim of gene selection is to reduce the dimensions (features) of gene expression data leading to eliminating irrelevant and redundant genes.

Methods

This paper presents a hybrid gene selection method based on graph theory and a many-objective particle swarm optimization (PSO) algorithm. To this end, a filter method is first utilized to reduce the initial space of the genes. Then, the gene space is represented as a graph to apply a graph clustering method to group the genes into several clusters. Moreover, the many-objective PSO algorithm is utilized to search an optimal subset of genes according to several criteria, which include classification error, node centrality, specificity, edge centrality, and the number of selected genes. A repair operator is proposed to cover the whole space of the genes and ensure that at least one gene is selected from each cluster. This leads to an increasement in the diversity of the selected genes.

Results

To evaluate the performance of the proposed method, extensive experiments are conducted based on seven datasets and two evaluation measures. In addition, three classifiers—Decision Tree (DT), Support Vector Machine (SVM), and K-Nearest Neighbors (KNN)—are utilized to compare the effectiveness of the proposed gene selection method with other state-of-the-art methods. The results of these experiments demonstrate that our proposed method not only achieves more accurate classification, but also selects fewer genes than other methods.

Conclusion

This study shows that the proposed multi-objective PSO algorithm simultaneously removes irrelevant and redundant features using several different criteria. Also, the use of the clustering algorithm and the repair operator has improved the performance of the proposed method by covering the whole space of the problem.

Peer Review reports

Background

In recent years, microarray technology has made feasible the simultaneous monitoring of thousands of genes. In particular, this type of technology has been successfully applied as a fruitful source of data in bioinformatics research in order to categorize, identify, and express thousands of genes in a wide range at the same time. Gene selection plays a crucial role in the diagnosis of various diseases. Gene expression databases tend to contain large volumes of features, but low numbers of samples. Such databases often include immense amounts of irrelevant and unnecessary attributes where only certain minute parts of genes are related to a given disease [1]. As a result, gene selection is an extremely important element in applying gene expression databases to disease diagnosis, effectively.

Gene selection methods can be classified into four categories: filter, wrapper, hybrid, and embedded [2, 3]. The filter method is related to the problems in which gene selection is carried out independently of any learning algorithm. In other words, gene selection is performed as a separate preprocessing task. A detailed and accurate statistical analysis of the selected genes is essential to carry out the gene selection method without the help of any learning model. The wrapper approach applies a defined learning algorithm in order to establish the efficiency and usefulness of the selected subsets. Wrappers are capable of producing much better results; however, they are much more costly to set up and often also involve analyzing a large number of genes. The hybrid approach combines the filter and wrapper methods and tries to exploit both of them. Finally, the embedded approach seeks to perform gene selection as a part of the learning process and is specific to a given learning sample.

A number of population-based evolutionary algorithms, including genetic algorithm (GA) [4,5,6,7,8], ant colony optimization (ACO) [9,10,11,12], tabu search (TS) [13, 14], simulated annealing (SA) [15, 16], and particle swarm optimization (PSO) [17,18,19,20] have attracted significant attention in the field of bioinformatics because of their ease of use and acceptable results in tackling the challenge of gene selection [21, 22]. However, most of these methods suffer from some common issues, including: (1) requiring a large amount of computational time, (2) an inability to produce acceptable outcomes because they do not focus sufficiently on reducing the size of the selected genes, and (3) the need to carry out a very large number of iterations and parameters in order to tune them. A further drawback of such population-based approaches is that they normally have only one main objective—either to increase the classification accuracy or to reduce the number of selected genes—and do not take sufficient account of other objectives, such as the strength of the relevancy between genes.

In order to tackle the above-mentioned issues, a gene selection method called MaPSOGS is proposed in this paper based on a many-objective PSO algorithm. Specifically, this paper proposes a hybrid filter-wrapper approach based on a many-objective PSO algorithm. In the proposed method, a graph clustering model is used to group the genes into several clusters. To this end, the initial set of genes are filtered using the Fisher score and then, the remaining genes are represented as a graph. Moreover, a novel operator called "repair operator" is proposed to improve the solutions in the PSO algorithm by selecting the genes from different clusters. This can make a more diverse subset of genes leading to an improvement in the performance of the classification methods. The proposed approach contains a number of contributions compared to other well-known approaches:

  1. 1.

    Although other approaches only consider the direct relevancy between genes, the proposed method uses the criterion of edge centrality to clearly measure the intensity of the relevancy between the genes.

  2. 2.

    The proposed method uses a many-objective PSO algorithm based on several important objectives for gene selection including: increasing the classification accuracy, reducing redundancy, reducing the rate of false positives, and minimizing the size of the final selected set of genes.

  3. 3.

    The proposed method applies a graph clustering algorithm to group similar genes together as a cluster and then uses a repair operator to ensure that the entire space of the genes is explored in the search process.

  4. 4.

    Determining the number of genes that should be selected is an important challenge in gene selection. This is due to the fact that the number of relevant genes is unknown; thus, the optimum number of genes to be selected is also unknown. However, in the proposed method, the optimum number of genes to be selected is measured automatically based on the overall structure of the original genes and the similarity between them.

In the literature, many studies have been conducted to develop gene selection models based on metaheuristic algorithms. A hybrid gene selection method is proposed in [20], which is based on an mRMR (minimum redundancy maximum relevance) filter. This method employs the mutual information between genes to select an optimal set of genes. In [21], a gene selection method called GANN is proposed based on the combination of genetic algorithm (GA) and a neural network model. Shreem et al. [23] proposed a gene selection method called R-m-GA, which combines the ReliefF, mRMR, and GA algorithms. To this end, a set of candidate genes is selected by applying the ReliefF algorithm. Then, the redundancy of the candidate genes is reduced by applying the mRMR algorithm. Finally, the genetic algorithm is used to select the final subset of genes using an IB1 classifier.

The particle swarm optimization (PSO) algorithm is used by many scholars to develop effective models for gene selection [24]. A feature selection method based on the univariate filter model and the PSO-based wrapper model was developed in [25]. Inbarani et al. [26] proposed a hybrid model that was adapted for medical applications. In their work, to improve disease diagnosis in medical datasets, a hybrid feature selection based on PSO and rough sets theory is applied. The PSO algorithm and the support vector machine (SVM) classifier were also integrated in [27] for feature selection and parameter optimization. Also, a distributed parallel architecture is used to overcome the high computational complexity in high dimensional datasets.

A gene selection method is introduced in [28] based on a recursive PSO algorithm. In this method, the dimensions of genes with large spaces are reduced in an iterative step. To this end, the filtering-oriented ranking approaches are applied with the recursive PSO algorithm to achieve an appropriate subset of genes. A hybrid gene selection method called IG-ISSO is proposed in [29] which is based on the combination of IG and improved simplified swarm optimization. In particular, the IG approach is applied to choose more effective genes, and then, the improved simplified swarm optimization approach is utilized to search for the optimum set of genes. In [18], the authors proposed a model to select genes based on the black hole embedding in BPSO algorithm. This model is able to increase the efficiency of the BPSO algorithm by improving both exploration and exploitation phases in the search process. In [30], the combination of the black hole algorithm and decision tree is used to propose an effective gene selection method. Chuang et al. [28] proposed a gene selection method that employs the combination of the BPSO, genetic algorithm, and KNN classifier. Sahu et al. [31] proposed a two-phases gene selection method. In the first phase, the initial genes are split into a number of clusters by using the K-MEANS algorithm. Then, the genes of each cluster are ranked using the SNR score, and the genes with the highest scores are selected as a new subset. In the second phase, the new produced subset of genes is used as the input of the PSO algorithm to obtain the final subset. Xi et al. [32] proposed a gene selection method based on the binary quantum-behaved particle swarm optimization (BQPSO) for cancer classification. In [33], a PSO algorithm called PPSO has been developed for gene selection. PPSO uses a new representation space to decrease the search space and also uses a new fitness function to better evaluate the solutions. In [34], an improved version of the PSO algorithm is introduced to improve the performance of the KNN classifier. To this end, the exploitation capability of the PSO algorithm is enhanced by determining the global optimal solution more efficiently. Banka et al. [35] proposed a feature selection method for high-dimensional data based on Hamming distance-based binary particle swarm optimization (HDBPSO). Hamming distance is used in their method as a proximity measure for updating the particle velocity during a binary PSO search process to select the requisite feature subsets. Another hybrid feature selection algorithm utilizing particle swarm optimization is proposed in [36]. This method, called HPSO-LS, selects the less correlated and salient feature subset by utilizing a new local search. Jain et al. [37], integrated the correlation feature selection with modified binary PSO algorithm for gene selection and cancer classification. This method eliminates irrelevant and redundant genes to choose a high-relevant subset. In [38], an approach to reduce dimensionality in a medical dataset was developed using PSO-based regression. Moreover, the Bayesian information criterion is combined with PSO and logistic regression as a fitness function.

There have been several studies that combine GA with PSO to benefit from both their advantages and cover their drawbacks. Li et al. [39] presented a gene selection method using a hybrid of PSO/GA and SVM as a classifier. A hybrid PSO/GA algorithm is proposed by [40] along with Artificial Neural Networks (ANNs) to uncover biomarkers from microarray data. In [41], BPSO and combat genetic algorithms (CGA) are used to reduce the number of genes in gene expression and achieve a low classification error rate.

Methods

Many objective optimization

Optimization problems can be divided into two categories based on the number of objective functions and optimization criteria: (1) single-objective optimization problems and (2) multi-objective optimization problems. In single-objective optimization problems, there is a target function with only one criterion in which the purpose is to find an optimal value to maximize or minimize this function. On the other hand, multi-objective optimization problems contain several objective functions with some conflicting criteria. The purpose of these problems is to find an optimal solution by considering all the conflicting criteria. Therefore, the optimization process in multi-objective problems is more critical than single-objective problems. A many-objective optimization problem is a multi-objective problem with at least four objective functions, which can be represented as follows [42]:

$$\begin{aligned} & minimize\; F\left( x \right) = \left( {f_{1} \left( x \right),f_{2} \left( x \right), \ldots ,f_{m} \left( x \right)} \right)^{T} , \\ & subject\, to\; x \in {\Omega } \\ \end{aligned}$$
(1)

where the decision vector \(x = \left( {x_{1} ,x_{2} , \ldots , x_{n} } \right)\) is a member of the nonempty decision space \({\Omega }\), and the objective function \(F\) consists of \(m \left( {m \ge 4} \right)\) objective functions.

The Pareto dominance strategy is mainly applied to evaluate the solutions of the many-objective optimization problems according to the Pareto optimal solution. The definitions of the Pareto dominance and Pareto optimal solution are represented in the following [42]:

Theorem 1: Pareto dominancePareto dominance.

[43] Given two solutions \(x , y \in {\Omega }_{f}\) and their corresponding objective vectors \(F\left( x \right), F\left( y \right) \in R^{m}\), \({\varvec{x}}\) dominates \({\varvec{y}}\) (denoted as \({\varvec{x}} < {\varvec{y}}\)) if and only if \(\forall i \in \left\{ {1, 2, \ldots ,m} \right\}, f_{i} \left( x \right) \le f_{i} \left( y \right)\) and \(\exists j \in \left\{ {1, 2, \ldots ,m} \right\}, f_{j} \left( x \right) \le f_{j} \left( y \right)\).

Theorem 2: Pareto optimal solution. A solution \(x^{*} \in {\Omega }_{f}\) is Pareto optimal if there is not any other solution \(x \in {\Omega }_{f}\) that dominates \(x^{*}\).

Proposed method

In this section, we aim to introduce our proposed gene selection method for medical diagnosis problems which is called MaPSOGS. Figure 1 indicates an overview of the proposed method. This method consists of four main steps: (1) filtering out of genes, (2) graph representation, (3) genes clustering, and (4) genes selection. In the first step of the proposed method, the Fisher score is used to filter out those genes whose Fisher score values are less than a threshold value. In the second step, a graph is represented in which the reduced set of genes are the nodes and their relationships are used to construct the edges. Then, a clustering approach is applied to the graph to cluster the genes into appropriate groups. Finally, a many-objective optimization method based on the PSO algorithm is employed to select the final set of genes according to different objective functions. The details of the main steps of the proposed method are discussed in the following subsections.

Fig. 1
figure 1

The Overview of the proposed method

Filtering out of genes

Real-world datasets mainly contain a large number of genes, while in most cases, only a very small number of them are appropriate and other genes are irrelevant or redundant. Moreover, processing large-scale datasets with enormous genes is very time-consuming. Therefore, in this step, we aim to provide an effective mechanism to filter out irrelevant or redundant genes from the original set of genes. This helps to ease the process of selecting genes using the proposed many-objective optimization mechanism. To do this, we calculate the Fisher score for each gene using the following equation:

$$Score\left( {G_{i} } \right) = \frac{{\mathop \sum \nolimits_{k = 1}^{C} n_{i} (\overline{g}_{i}^{k} - \overline{g}_{i} )^{2} }}{{\mathop \sum \nolimits_{k = 1}^{C} n_{i} (\sigma_{i}^{k} )^{2} }}$$
(2)

where, \(C\) is the number of classes in the dataset, \(n_{i}\) is the number of samples in class \(i\),\(\overline{g}_{i}\) denotes the mean value of all patterns related to the gene \(G_{i}\), and \({\overline{\text{g}}}_{{\text{i}}}^{{\text{k}}}\) and \(\sigma_{i}^{k}\) denote mean and variance of class \(k{ }\) related to the gene \(G_{i}\). After calculating the Fisher score for all genes, a subset of them is selected using a threshold value named \(\lambda\). In other words, those genes whose Fisher scores are lower than \(\lambda\) are filtered out from the original set, and other genes are considered the selected genes. It is worth noting that the selected subset of genes is used in the next steps of the proposed method.

Graph representation

In this step, the space of genes is represented as a graph to be used in the next step of the proposed method (i.e., genes clustering). To this end, the set of genes is mapped to a weighted graph \(G = \left( {Ge, E, w_{Ge} } \right)\) in which \(Ge = \left\{ {G_{1} , G_{2} , \ldots , G_{n} } \right\}\) is the set of genes, \(E = \left\{ {\left( {G_{i} ,G_{j} } \right): G_{i} , G_{j} \in Ge} \right\}\) denotes the edges of the graph, and \(w_{Ge} :\left( {G_{i} , G_{j} } \right) \to {\mathbb{R}}\) is a function representing the similarity values between the genes. It should be noted that in the graph representation model, the genes are considered as the nodes and the similarity values between the genes are used as the weights of edges in the graph. There are several approaches to calculate the similarity values between genes. Among these approaches, the Euclidean distance, the cosine similarity, and the Pearson correlation coefficient are the most popular functions used to calculate the similarity values between genes. In the proposed method, we use the Pearson correlation coefficient due to its efficiency in calculating the similarity value between two input vectors. The Pearson correlation coefficient between two genes \(G_{i}\) and \(G_{j}\) is calculated as follows:

$$w_{ij} = \left| {\frac{{\mathop \sum \nolimits_{p} \left( {x_{i} - \overline{{x_{i} }} } \right)\left( {x_{j} - \overline{{x_{j} }} } \right)}}{{ \sqrt {\mathop \sum \nolimits_{p} \left( {x_{i} - \overline{{x_{i} }} } \right)^{2} } \sqrt {\mathop \sum \nolimits_{p} \left( {x_{j} - \overline{{x_{j} }} } \right)^{2} } }}} \right|$$
(3)

where \(x_{i}\) and \(x_{j}\) represent the vectors of the genes \(G_{i}\) and \(G_{j}\), respectively. Moreover, \(\overline{{x_{i} }}\) and \(\overline{{x_{j} }}\) denote the mean values of the vectors \(x_{i}\) and \(x_{j}\) between \(p\), respectively.

In the proposed method, the weights between the genes in the graph are normalized using a nonlinear scaling function to improve the effectiveness of the proposed method. Therefore, the weights of the graph are mapped to a new range [0, 1]. To this end, the weights of the graph (the similarity values between genes) are normalized using the following equation [44]:

$$\hat{w}_{ij} = \frac{1}{{1 + {\text{exp}}( - \frac{{w_{ij} - \overline{w}}}{\sigma })}}$$
(4)

where \(w_{ij}\) is the similarity value between the genes \(G_{i}\) and \(G_{j}\), \(\overline{w}\) and \(\sigma\) are the mean and standard deviation of all calculated similarity values, respectively.

Genes clustering

In this step, a graph clustering approach is used to cluster the genes into appropriate clusters. The main purpose of this step is to split up the initial genes into a number of clusters according to their similarity values. Therefore, similar genes can be assigned to the same cluster while their similarity values with the genes of other clusters will be low. In the proposed method, we use the Louvain algorithm [45] as a graph clustering approach. It should be noted that this graph clustering approach is applied to the graph which is constructed in the graph representation step. To improve the efficiency of the clustering approach, the initial graph is updated by removing the edges whose weights are less than the threshold \(\theta\). The main idea of the Louvain algorithm is to perform a fast and efficient mechanism to detect the communities (clusters) by maximizing the modularity function. To this end, at first, each node is considered as a cluster, and then the clustering procedure is done in two iterative steps as follows:

Step 1: For each node \(i\), the benefit of assigning this node to cluster C is calculated using the following equation:

$$\Delta Q = \left[ {\frac{{\mathop \sum \nolimits_{in} + k_{i,in} }}{2m} - \left( {\frac{{\mathop \sum \nolimits_{tot} + k_{i} }}{2m}} \right)^{2} } \right] - \left[ {\frac{{\mathop \sum \nolimits_{in} }}{2m} - \left( {\frac{{\mathop \sum \nolimits_{tot} }}{2m}} \right)^{2} - \left( {\frac{{k_{i} }}{2m}} \right)^{2} } \right]$$
(5)

where \(\mathop \sum \limits_{in}\) is the total weights in cluster \(C\), \(\mathop \sum \limits_{tot}\) is the total weights of the edges that are connected to the nodes of cluster \(C\), \(k_{i}\) is the degree of node \(i\), \(k_{i, in}\) indicates the sum of the weights of the edges between node \(i\) and other nodes in its community, and \(m\) is the sum of the weights of all edges in the graph.

Step 2: Each node is assigned to a cluster that maximizes the modularity function. Therefore, new clusters can be obtained based on this strategy.

These two steps are repeated until the structure of clusters is no longer changed. The output of this step is a number of clusters where each cluster contains a number of similar genes. These clusters are used in the next step of the proposed method to obtain the final set of genes. The Louvain algorithm is a simple and efficient method for identifying clusters in large graphs. The computational complexity of this algorithm is \(O\left( {n\log n} \right)\), where \(n\) is the number of nodes.

Gene selection

In this step, the final subset of genes is selected using the many-objective PSO algorithm. To do this, two important issues must be considered: (1) the representation of solutions and (2) the definition of fitness function. In the proposed method, each particle represents a solution whose length is equal to the total number of genes. If a gene is selected, the value of its position in the solution will be 1, otherwise; it will be 0. Moreover, the fitness function is defined based on several criteria that must be minimized using the optimization process.

In the many-objective PSO algorithm, first of all, the initial population of the particles is randomly generated. Then, the fitness value of each particle is calculated using the defined fitness function. It is worth noting that the fitness function is defined based on several criteria: classification error, edge centrality, node centrality, number of selected genes, and specificity metric. These criteria are considered as the objective functions that should be either maximized or minimized. Classification error evaluates the rate of the samples that are not correctly classified, which is calculated as follows:

$$Classification \,Error = \frac{FP + FN}{{TP + FP + TN + FN}}$$
(6)

where \(FP, FN,TP, and TN\) are respectively the numbers of false positive, false negative, true positive, and true negative.

The edge centrality is a criterion for evaluating the intensity of relevancy between genes. One of the purposes of the proposed method is to select a set of genes with the least redundancy. To this end, edge centrality [46] has been applied to measure the intensity of relevancy between genes which should be minimized. The node centrality [47] is used as a criterion for assessing the influence of nodes in the graph. The higher centrality of a node leads to make a better representative for other neighboring nodes. Besides, since the true negative rate of results is highly important for medical diagnosis problems, the specificity criterion is applied to measure the ratio of correctly identified negative cases which should be maximized. The specificity metric is calculated as follows:

$$Specificity\left( {TNR} \right) = \frac{TN}{{TN + FP}}$$
(7)

The number of selected genes is another criterion that should be minimized. This is due to the fact that the proposed method aims to select a lower number of genes as the final subset of genes. According to the considered criteria, the many-objective fitness function for the subset of genes \(GS\) is defined as follows:

$$minimize F\left( {GS} \right) = \left( {f_{1} \left( {GS} \right),f_{2} \left( {GS} \right),f_{3} \left( {GS} \right),f_{4} \left( {GS} \right),f_{5} \left( {GS} \right)} \right)^{T}$$
(8)

where

$$f_{1} \left( {GS} \right) = Classification\, Error$$
(9)
$$f_{2} \left( {GS} \right) = \mathop \sum \limits_{{G_{i} ,G_{j} \in G}} EC\left( {G_{i} ,G_{j} } \right); EC\left( {G_{i} ,G_{j} } \right) \,is\, the\, edge\,centrality\,between\,gene\,pair\, of\, G_{i}$$
(10)
$$f_{3} \left( {GS} \right) = \frac{1}{{NC\left( {GS} \right)}} ;\; NC\left( {GS} \right)\, is\, the\, sum\, of\, the\, node\, centrality\, of\, all\, genes\, in\, GS$$
(11)
$$f_{4} \left( {GS} \right) = \frac{1}{{Specificity\left( {TNR} \right)}}$$
(12)
$$f_{5} \left( {GS} \right) = \left| {GS} \right|;\; \left| {GS} \right|\, is\,the\,number\,of\, genes\, in\, GS$$
(13)

After calculating the fitness values of particles, their positions are updated based on an effective mechanism. To this end, the position of each particle is updated using its best position and also the global best position that are achieved in the previous iterations. In this regard, the position of particle \(i\) is denoted by vector \(x_{i}\) and also the velocity vector of this particle is defined as \(v_{i}\). Then, the position of each particle \(i\) is updated using the following equations [48]:

$$v_{i} \left( {t + 1} \right) = wv_{i} \left( t \right) + c_{1} r_{1} \left( {x_{Best,i} \left( t \right) - x_{i} \left( t \right)} \right) + c_{2} r_{2} (g_{best,i} \left( t \right) - x_{i} \left( t \right))$$
(14)
$$x_{i} \left( {t + 1} \right) = x_{i} \left( t \right) + v_{i} \left( t \right).t$$
(15)

where \(x_{Best,i}\) and \(g_{best}\) denote the best position of the particle \(i\) and the best global position, respectively. Moreover, \(w\) is inertia weight, \(c_{1}\) and  \(c_{2}\) are two positive constants, and \(r_{1}\) and \(r_{2}\) are two random values in the range \(\left[ {0,{ }1} \right]\). It should be noted that the value of parameter \(w\) is set to \(w \approx \left[ {0.5,{ }0.9} \right]\) in the proposed method.

In the proposed method, a repair operator is introduced which is applied to the particles in order to readjust the number of the selected genes from each cluster. To this end, we use the clusters of genes that are obtained in the previous step of the proposed method. The proposed repair operator determines that which genes should be selected or removed according to a criterion. For this purpose, at the beginning of the PSO algorithm, the effectiveness of each gene is calculated using the Fisher score (Eq. (2)). Then, the obtained scores are mapped to a new range [0, 1] to use as the probability of selection or deletion of each gene in the repair process. The main purpose of the proposed repair operator is to increase the diversification of the selected genes. To this end, it tries to select the genes from all clusters uniformly, instead of selecting the genes from a small number of clusters. Therefore, if the number of selected genes from a cluster is less than \(\omega\), a number of genes with the lowest Fisher score in the selected subset will be replaced with the genes in this cluster with the highest fisher score. The main advantage of the proposed repair operator is to enhance the diversity of selected genes leading to an improvement in the performance of the proposed gene selection method. Figure 2 illustrates the overall schema of the proposed repair operator.

Fig. 2
figure 2

The overall schema of the proposed repair operator for a small example with ten genes. The number of clusters is set to three. The initial set of selected genes is \({\varvec{G}} = \left\{ {{\varvec{G}}_{1} ,\user2{ G}_{3} ,\user2{ G}_{7} ,\user2{ G}_{10} \user2{ }} \right\}\) that all of them belong to cluster 1 and 2 and no gene has been selected from cluster 3. After performing the repair operator, a gene with the lowest Fisher score (G7) is removed from the initial set of selected genes, and a gene from cluster 3 with the highest Fisher score (G9) is added to the final set of selected genes

The proposed many-objective PSO algorithm performs iteratively until the stopping criteria are satisfied. The stopping criteria are satisfied if the maximum number of iterations is reached or there is no change in the positions of particles. Finally, the best particle obtained by the many-objective PSO algorithm is considered as the final solution, which contains the genes selected as the output of the proposed gene selection method. The pseudo-code of the proposed gene selection method is represented in Algorithm 1.

figure a

Computational complexity analysis

In this section, the computational complexity of the proposed method is calculated. In the first step, the relevance values of the genes are evaluated using the Fisher measure. The computational complexity of this step is \(O\left( {ncp} \right)\) where \(n{ }\) is the number of genes in the original set, \(c\) is the number of classes, and \(p\) denotes the number of samples. The second step of the proposed method represents the gene space as a graph which its complexity is \(O\left( {m^{2} p} \right)\) where \(m\) is the number of the genes after applying the Filter method. In the third step, the Louvain community detection algorithm is used to cluster the genes into several clusters which is performed with the complexity of \(O\left( {m\log m} \right)\). Then, the proposed many-objective PSO algorithm is used to select the final set of genes. In this step, the fitness value of each particle should be calculated. The computational complexity of the SVM classification algorithm is \(O\left( {p^{2} .S} \right){ }\) where \(S = \omega {*}k\). Moreover, the computational complexity of calculating the edge centrality and node centrality is \(O\left( {m^{2} } \right)\). Therefore, the calculation of the fitness function has a computational complexity \(O\left( {I.A.p^{2} .S + I.A.m^{2} } \right)\) where \(I\) represents the maximum number of iterations and \(A\) is the number of particles. Also, in each iteration of the algorithm, it is necessary to apply the repair operator on all particles. The complexity of the repair operator is \({ }O\left( {I.A.k.m} \right)\), where \(k\) represents the number of clusters. Therefore, the computational complexity of this step of the proposed method is \({ }O\left( {I.A.k.m + I.A.p^{2} .S + I.A.m^{2} } \right)\), which can be reduced to \({ }O\left( {m^{2} + P^{2} .S} \right)\). According to the computational complexities of all steps, the final computational complexity of the proposed method will be \({ }O\left( {n.c.p + m^{2} .p + m.\log m + m^{2} + p^{2} .S} \right)\) which is reduced to \(O\left( {m^{2} .p + p^{2} .S} \right)\).

Results

To evaluate the performance of the proposed gene selection method, several experiments are performed, and their results are discussed in this section. These experiments are conducted on seven gene expression datasets that are gathered from http://www.biolab.si/supp/bi-cancer/projections/info/leukemia.html. The characteristics of these datasets are shown in Table 1. Moreover, the accuracy of classification methods is measured based on the selected set of genes and also the number of selected genes. In other words, the higher accuracy value and the lower number of selected genes show the higher performance for the gene selection models. To this end, different classification methods are employed, including Support Vector Machine (SVM), Decision Tree (DT), and K-Nearest Neighbors (KNN). Furthermore, the Weka software [49] is applied to build classifiers in the experiments. The proposed method is compared with several state-of-the-art gene selection models that are briefly described in the following:

  • Geometric particle swarm optimization (GPSO) [50]: a gene selection method based on the PSO algorithm that applies binary representation in Heming space.

  • PSO: a gene selection method which obtains a subset of genes using the basic version of the PSO algorithm and evaluates the selected genes by means of an SVM classifier.

  • RMA [51]: a gene selection method based on the recursive Memetic Algorithm (MA). The recursive strategy of this algorithm is used to make a balance between the exploitation and exploration phases of the search process.

  • RPSW [52]: a gene selection method based on the combination of the Return PSO Algorithm and different filtering ranking methods. It also applies a linear SVM weight vector to select the primary gene subset.

  • IG-ISSO [29]: a hybrid method which uses information gain (IG) as the filtering method to select a subset of genes that is most informative based on the amount of IG. Moreover, it uses an improved simplified swarm optimization (ISSO) algorithm to perform the search process.

  • Hybrid BPSO-BBHA [21]: a gene selection method which incorporates the Binary Black Hole Algorithm (BBHA) in the BPSO algorithm to facilitate and improve the efficiency of the BPSO.

  • PSOC4.5 [53]: a gene selection method based on the PSO algorithm and decision tree classifier. It uses the search capabilities of the PSO algorithm to interpret the knowledge of C4.5.

Table 1 The characteristics of the used datasets

Performance comparison

In this subsection, the experimental results are reported in terms of the number of selected genes and the average classification accuracy over ten independent runs for different gene selection methods. It is worth mentioning that a training set is considered to pick the final subset of genes. Also, a test set is applied to evaluate the selected genes. Table 2 shows the performance of different gene selection methods based on the number of genes which are selected by these methods. As we can see from this table, the proposed method achieves better results than the other models in most cases for the used datasets. The average number of selected genes for the proposed method is the best value among all the compared methods. Therefore, it can be concluded that the proposed gene selection method can significantly reduce the initial space of the genes by selecting the lower number of them than the other compared models. The average number of selected genes for the proposed method is 5.14 while the second-best value is obtained by the RMA model which is equal to 5.42.

Table 2 Performance comparison of different gene selection methods based on the number of selected genes

The performance of the compared gene selection methods is also evaluated based on the accuracy of classification models which are applied to the selected subset of genes. To this end, the leave one out cross validation approach is used to measure the accuracy of classification models. In these experiments, we use three different classification models including SVM, DT, and KNN classifiers. The results of experiments are reported for SVM, DT, and KNN classifiers in Tables 3, 4, and 5, respectively. As it can be seen from these results, the proposed method often has higher classification accuracy than the other methods in most cases. In addition, the standard deviation value and the rank of each method are reported in these tables. The results demonstrate that the proposed method achieves the first rank in comparison to other gene selection models based on all used datasets and all considered classifiers. Table 3 shows that in all cases except for the Colon dataset, the proposed method has the best performance among all the other gene selection methods. On the other hand, while the RMA approach has the best performance for the Colon dataset, the proposed method takes the second ranking position.

Table 3 The results of different gene selection methods over different datasets based on SVM classifier
Table 4 The results of different gene selection methods over different datasets based on DT classifier
Table 5 The results of different gene selection methods over different datasets based on KNN classifier

We report the classification accuracy for SVM, DT, and KNN classifiers in Tables 3, 4, and 5. As you can see from these numerical results, the proposed method is superior to other swarm intelligence-based gene selection methods according to all datasets. For example, for the MLL dataset, MaPSOGS obtains a 96.38% classification accuracy while for Geometric PSO, PSO, RMA, RPSW, EPSO, Hybrid BPSO-BBHA, and PSOC4.5, this value was reported 81.26%, 81.67%, 87.19%, 91.23%, 86.53%, 85.98%, 86.75%, and 91.63%, respectively. In addition, the results of these tables show that the proposed method has an average ACC of 98.68%, 95.65%, and 94.59% for the SVM, DT, and KNN classifiers, respectively. Compared to the second-ranked RMA method which has the average ACC of 96.04%, 93.38%, and 94.59%, these ACCs demonstrate improvements of 2.64%, 2.27%, and 1.66% for the proposed method. Also, the results of Table 2 indicate that the MaPSOGS method selects fewer genes than other methods. The average number of genes selected by the proposed method is 5.14.

Sensitivity analysis of the parameters

In the second step (genes clustering) of the proposed method, the edges with associated weights lower than the parameter \(\theta\) will be removed from the graph. The performance of the genes clustering algorithm depends on the value of the parameter \(\theta\). The value of this parameter can be set to any value in the range [0,1]. If \(\theta\) is set to a small value, more edges will be considered in the graph clustering algorithm and the number of obtained clusters will be declined. On the other hand, if \(\theta\) is set to a high value, the graph clustering algorithm identifies a greater number of clusters. In this section, a series of experiments are conducted to analyze the effect of parameter \(\theta\) on the performance of the genes clustering algorithm. Figure 3 shows the accuracy of the proposed method based on SVM classifier in terms of different values of the parameter \(\theta\) for AMLGSE2191, Colon, DLBCL, and Leukemia datasets. Moreover, the number of obtained clusters is also reported in Fig. 3. These results indicate that the higher value of the parameter \(\theta\) leads to an increasement in the number of obtained clusters. On the other hand, the accuracy of the proposed method is declined when the value of the parameter \(\theta\) exceeds a specific value. For instance, when the value of the parameter \(\theta\) exceeds 0.6, the accuracy of the proposed method will be reduced for the AMLGSE2191 dataset. It can be concluded from these results that when the parameter \(\theta\) is set to a small value, the graph clustering algorithm identifies lower number of clusters. As a result, in this case, the proposed method selects the smaller number of genes and thus, most representative genes cannot be selected to reduce the classifier accuracy.

Fig. 3
figure 3

Average classification accuracy and the number of obtained clusters for different values of the parameter \({\varvec{\theta}}\) based on SVM classifier: a AMLGSE2191 dataset, b Colon dataset, c DLBCL dataset, and d Leukemia dataset

To investigate the convergence speed of the proposed method, the convergences of MaPSOGS, RMA, BPSO-BBHA, EPSO, RPSW, Geometric PSO, PSOC4.5, and PSO methods as well as MaPSOGS algorithm without repair operator are compared based on Colon and DLBCL datasets. Figure 4 shows the convergence speed of these methods in terms of different iterations. Figure 4 (a) shows the convergence speed for the Colon dataset. As it can be seen from this figure, the convergence speed of the proposed method is faster than other ten models. Moreover, the classification error of the proposed method is lower than the other models in most cases. The superiority of the proposed method in respect to the model without the repair operator demonstrates that the repair operator has a positive effect on the accuracy of classification models. Figure 4 (b) shows the convergence speed of the compared methods on the DLBCL dataset. These results also show that the proposed method has less classification error and a faster convergence speed in comparison to other models.

Fig. 4
figure 4

Comparison of convergence speed of MaPSOGS, RMA, MaPSOGS (without repair operator), Hybrid BPSO-BBHA, EPSO, RPSW, Geometric PSO, PSOC4.5 and PSO models based on: a the Colon dataset and b the DLBCL dataset

Statistical analysis of the proposed method

In order to have a statistical analysis on the results of the proposed method and other compared methods, the Friedman test [54] is conducted. This statistical test is a nonparametric test which is used to compare the performance of different methods based on different medical datasets. In this way, the rank of each gene selection model can be obtained according to the used datasets. To this end, we use the SPSS statistics acquired by IBM [55]. In the Friedman test, hypothesis \(\mathrm{H}0\) is based on the sameness of the average ranks among the groups. Rejection of the null hypothesis means that at least two groups have significant differences. In the analysis of the Friedman test results, it is impossible to determine that if the level of significance is less than the level of error, the difference between at least a pair of specimens is deducted. Since the test error is considered 5%, the level of significance must be lower than 0.05 to satisfy the constraint. Tables 6 and 7 report the Friedman test results for the proposed method compared with the other gene selection methods. In particular, Table 6Table shows the average rankings of all compared models based on different classification methods. As we can see from these results, the proposed method has the best ranking among all compared models. Therefore, it can be concluded that the proposed method is the best performer according to all considered classifiers. Table 7 shows that the Friedman test has reported a P value of 0.0018 for the classification accuracy of the SVM classifier. Since this is below 0.05, we can claim that the results of the proposed method are significantly different from those of other methods. Other P values achieved by the Friedman test for DT and KNN classifiers confirm the claim that the proposed method is significantly different from other compared methods.

Table 6 Average ranks of different methods on SVM, DT, and KNN classifiers
Table 7 The results of the statistical test

Discussion

This section briefly explains why the performance of the proposed MaPSOGS method is better than other compared gene selection methods. The superiority of the proposed method is due to three main contributions that are discussed in the following.

  1. 1.

    Irrelevant and redundant genes severely affect the accuracy of the learning algorithms [56,57,58]. Thus, gene selection methods should be able to identify and remove these irrelevant and redundant genes in the original space. Most of the existing gene selection methods can effectively remove the irrelevant genes but fail to handle the redundant genes. In the univariate methods (i.e., L-Score, F-Score, and RelifF), the relevance of a gene with others is individually measured and the possible dependency between the genes will be ignored in the gene selection process. Thus, these methods cannot be able to remove the redundant genes precisely. On the other hand, most of the multivariate gene selection methods only eliminate the redundant genes without paying attention to the irrelevant genes. In this paper, we develop a novel gene selection method which can efficiently deal with both irrelevant and redundant genes. The proposed method selects the genes with minimum similarity with those of the previous selected ones using the edge centrality measure while maximizes the dependency with the target class using the error classification rate. By applying these two objective functions, the redundant and irrelevant genes have a lower probability of being selected.

  2. 2.

    One of the main shortcomings of existing gene selection methods is to choose the genes independently and consider the direct similarity between genes. To address this issue, the proposed method considers the intensity of relevancy between genes using the edge centrality criterion. Also, the specificity criterion is used in the proposed method to identify the negative samples leading to an improvement in the accuracy of classification models.

  3. 3.

    The proposed method uses a graph clustering model to cluster the genes into appropriate clusters. This leads to group the similar genes into the same cluster in which the members of each cluster are as far as possible to other clusters. Then, a repair operator is proposed to guarantee that the selected genes belong to different clusters. This operator could significantly increase the diversity of the selected genes leading to enhance the efficiency of the proposed gene selection method. Also, the repair operator increases the exploration capability of the proposed method and thus, improves the ability of the proposed method in finding the global optimal subset of the genes.

  4. 4.

    In order to select the optimal number of genes for the medical diagnosis, a reliable gene selection method should identify the optimal number of genes. When there are too many genes selected in a gene selection method, it increases the likelihood of selecting redundant and irrelevant genes, resulting in a reduced prediction accuracy. In contrast, too few genes may not be able to represent all of the original information. In this study, a many-objective fitness function is proposed, taking into account five main objectives, including the relevance of the gene, the redundancy of the gene, the classification error, specificity and the size of the gene subset. Therefore, redundant and irrelevant genes are selected with a low probability. Also, the selected genes provide sufficient information for diagnosis and prediction tasks. However, it is necessary to specify the number of genes in advance. The number of useful genes is hard to estimate before gene selection in actual medical applications. Whilst, the proposed method automatically determines the number of final gene set.

  5. 5.

    Swarm intelligence methods require consideration of the conflicting goals of exploring the search space and exploiting the best solutions. Exploration encompasses the process of generating various solutions so as to gain a global perspective on the search space. On the other hand, exploitation involves concentrating the search in a good region. A good balance between these two objectives will improve the performance of the searching method. By employing MaPSOGS, we aim to show that there is a powerful gene selection method that can balance exploitation and exploration capabilities while achieving faster convergence speed. We developed two strategies to balance between initial population generation and mutation operators in this paper. Accordingly, the node-centrality is used to enhance the exploitation in the proposed method. On the other hand, repair operator is considered as a diversification operator to avoid being stuck in local optimal by encouraging search exploration.

Conclusion

An effective gene selection method is proposed in this paper which is based on a many-objective PSO algorithm, the strategies of the graph theory, and also graph clustering. To this end, a filter approach is used to reduce the number of genes in the original set and then, a graph clustering algorithm is applied to the graph representing the genes and their relationships to cluster the genes into appropriate clusters. Moreover, a many-objective PSO algorithm is used to obtain an optimal subset of genes according to an SVM classifier as an evaluator. To define the fitness function of the many-objective PSO algorithm, different criteria are considered including classification error, node centrality, specificity, edge centrality, and also the number of genes selected by the proposed method. In addition, a repair operator is proposed to increase the diversity of the selected genes leading to enhance the performance of the proposed method. The proposed method is compared with seven well-known gene selection methods based on different medical datasets. The experimental results demonstrated that the proposed gene selection method can significantly outperform other compared models. Several user-specified parameters are used in the proposed method which their appropriate values should be determined initially. Gene selection algorithms rely on these parameters because they control how the learning model behaves and have a considerable impact on the final prediction. In order to select the best parameter values, repeating the process of setting parameters, generating several predictions with different combinations of values, and then testing the results to pick the most accurate ones is necessary. Therefore, it is necessary to optimize the parameters to obtain the best results. Exhaustive search algorithms can be used to optimize parameters’ values. In future work, an optimization method can be used to adjust the parameters. Also, social network analysis techniques such as popularity and community detection can be used to select the final gene set.

Availability of data and materials

The datasets used during the current study are freely available at http://www.biolab.si/supp/bi-cancer/projections/info/leukemia.html.

References

  1. S. Vanjimalar, D. Ramyachitra, P. Manikandan. A review on feature selection techniques for gene expression data. In: 2018 IEEE International Conference on Computational Intelligence and Computing Research (ICCIC). 2018. p. 1–4.

  2. Saeys Y, Inza I, Larranaga P. A review of feature selection techniques in bioinformatics. Bioinformatics (Oxford, England). 2007;23:2507–17.

    Article  CAS  Google Scholar 

  3. George V, Velanganny C. Review on feature selection techniques and the impact of Svm for cancer classification using gene expression profile. Int J Comput Sci Eng Surv. 2011;2:16–27.

    Article  Google Scholar 

  4. Zhu Z, Ong Y-S, Dash M. Markov blanket-embedded genetic algorithm for gene selection. Pattern Recognit. 2007;40(11):3236–48.

    Article  Google Scholar 

  5. Li S, Wu X, Hu X. Gene selection using genetic algorithm and support vectors machines. Soft Comput. 2008;12(7):693–8.

    Article  CAS  Google Scholar 

  6. Bonilla Huerta E, Duval B, Hao J-K. A hybrid LDA and genetic algorithm for gene selection and classification of microarray data. Neurocomputing. 2010;73(13):2375–83.

    Article  Google Scholar 

  7. Adie EA, Adams RR, Evans KL, Porteous DJ, Pickard BS. Speeding disease gene discovery by sequence based candidate prioritization. BMC Bioinform. 2005;6:55–55.

    Article  Google Scholar 

  8. Rostami M, Berahmand K, Forouzandeh S. A novel community detection based genetic algorithm for feature selection. J Big Data. 2021;8(1):2.

    Article  Google Scholar 

  9. Yu H, Gu G, Liu H, Shen J, Zhao J. A modified ant colony optimization algorithm for tumor marker gene selection. Genomics Proteomics Bioinform. 2009;7(4):200–8.

    Article  CAS  Google Scholar 

  10. Tabakhi S, Najafi A, Ranjbar R, Moradi P. Gene selection for microarray data classification using a novel ant colony optimization. Neurocomputing. 2015;168:1024–36.

    Article  Google Scholar 

  11. Vafaee Sharbaf F, Mosafer S, Moattar MH. A hybrid gene selection approach for microarray data classification using cellular learning automata and ant colony optimization. Genomics. 2016;107(6):231–8.

    Article  PubMed  CAS  Google Scholar 

  12. Sun L, Kong X, Xu J, Xue Z, Zhai R, Zhang S. A hybrid gene selection method based on reliefF and ant colony optimization algorithm for tumor classification. Sci Rep. 2019;9(1):8978.

    Article  PubMed  Google Scholar 

  13. Zhang H, Sun G. Feature selection using tabu search method. Pattern Recognit. 2002;35(3):701–11.

    Article  Google Scholar 

  14. Shen Q, Shi W-M, Kong W. Hybrid particle swarm optimization and tabu search approach for selecting genes for tumor classification using gene expression data. Comput Biol Chem. 2008;32(1):53–60.

    Article  CAS  Google Scholar 

  15. Filippone M, Masulli F, Rovetta S. Simulated annealing for supervised gene selection. Soft Comput. 2011;15(8):1471–82.

    Article  Google Scholar 

  16. M. Filippone, F. Masulli, S. Rovetta. Supervised classification and gene selection using simulated annealing. In: The 2006 IEEE International Joint Conference on Neural Network Proceedings. 2006. p. 3566–3571.

  17. Mohamad MS, Omatu S, Deris S, Yoshioka M. Particle swarm optimization for gene selection in classifying cancer classes. Artif Life Robot. 2009;14(1):16–9.

    Article  Google Scholar 

  18. Han F, Tang D, Sun Y-W-T, Cheng Z, Jiang J, Li Q-W. A hybrid gene selection method based on gene scoring strategy and improved particle swarm optimization (in Eng). BMC Bioinform. 2019;20(Suppl 8):289–289.

    Article  Google Scholar 

  19. Han F, et al. A gene selection method for microarray data based on binary PSO encoding gene-to-class sensitivity information. IEEE/ACM Trans Comput Biol Bioinform. 2017;14(1):85–96.

    Article  PubMed  Google Scholar 

  20. Rostami M, Forouzandeh S, Berahmand K, Soltani M. Integration of multi-objective PSO based feature selection and node centrality for medical datasets. Genomics. 2020;112(6):4370–84.

    Article  PubMed  CAS  Google Scholar 

  21. Pashaei E, Pashaei E, Aydin N. Gene selection using hybrid binary black hole algorithm and modified binary particle swarm optimization. Genomics. 2019;111(4):669–86.

    Article  PubMed  CAS  Google Scholar 

  22. Han F, Tang D, Sun Y-W-T, Cheng Z, Jiang J, Li Q-W. A hybrid gene selection method based on gene scoring strategy and improved particle swarm optimization. BMC Bioinform. 2019;20(8):289.

    Article  Google Scholar 

  23. Shreem S, Sheikh Abdullah S, Nazri MZA, Alzaqebah M. Hybridizing relief, mRMR filters and GA wrapper approaches for gene selection. J Theor Appl Inf Technol. 2012;46:1034–9.

    Google Scholar 

  24. Rostami M, Berahmand K, Nasiri E, Forouzandeh S. Review of swarm intelligence-based feature selection methods. Eng Appl Artif Intell. 2021;100:104210.

    Article  Google Scholar 

  25. Unler A, Murat A, Chinnam RB. mr2PSO: a maximum relevance minimum redundancy feature selection method based on swarm intelligence for support vector machine classification. Inf Sci. 2011;181(20):4625–41.

    Article  Google Scholar 

  26. Inbarani HH, Azar AT, Jothi G. Supervised hybrid feature selection based on PSO and rough sets for medical diagnosis (in Eng). Comput Methods Programs Biomed. 2014;113(1):175–85.

    Article  PubMed  Google Scholar 

  27. Huang C-L, Dun J-F. A distributed PSO–SVM hybrid system with feature selection and parameter optimization. Appl Soft Comput. 2008;8(4):1381–91.

    Article  Google Scholar 

  28. Chuang L-Y, Yang C-H, Li J-C, Yang C-H. A hybrid BPSO-CGA approach for gene selection and classification of microarray data. J Comput Biol. 2011;19:68–82.

    Article  PubMed  Google Scholar 

  29. Lai C-M, Yeh W-C, Chang C-Y. Gene selection using information gain and improved simplified swarm optimization. Neurocomputing. 2016;218:331–8.

    Article  Google Scholar 

  30. E. Pashaei, M. Ozen, N. Aydin. An application of black hole algorithm and decision tree for medical problem. In: 2015 IEEE 15th International Conference on Bioinformatics and Bioengineering (BIBE). 2015. p. 1–6.

  31. Sahu B, Mishra D. A novel feature selection algorithm using particle swarm optimization for cancer microarray data. Procedia Eng. 2012;38:27–31.

    Article  Google Scholar 

  32. Xi M, Juan L, Liu L, Fan F, Wu X. Cancer feature selection and classification using a binary quantum-behaved particle swarm optimization and support vector machine. Comput Math Methods Med. 2016;2016:1–9.

    Article  Google Scholar 

  33. Tran B, Xue B, Zhang M. A new representation in PSO for discretization-based feature selection. IEEE Trans Cybern. 2018;48(6):1733–46.

    Article  PubMed  Google Scholar 

  34. Chuang L-Y, Chang H-W, Tu C-J, Yang C-H. Improved binary PSO for feature selection using gene expression data. Computat Biol Chem. 2008;32(1):29–38.

    Article  CAS  Google Scholar 

  35. Banka H, Dara S. A Hamming distance based binary particle swarm optimization (HDBPSO) algorithm for high dimensionyal feature selection, classification and validation. Pattern Recognit Lett. 2015;52:94–100.

    Article  Google Scholar 

  36. Moradi P, Gholampour M. A hybrid particle swarm optimization for feature subset selection by integrating a novel local search strategy. Appl Soft Comput. 2016;43:117–30.

    Article  Google Scholar 

  37. Jain I, Jain VK, Jain R. Correlation feature selection based improved-Binary Particle Swarm Optimization for gene selection and cancer classification. Appl Soft Comput. 2018;62:203–15.

    Article  Google Scholar 

  38. Qasim OS, Algamal ZY. Feature selection using particle swarm optimization-based logistic regression model. Chemom Intell Lab Syst. 2018;182:41–6.

    Article  CAS  Google Scholar 

  39. Li S, Wu X, Tan M. Gene selection using hybrid particle swarm optimization and genetic algorithm. Soft Comput. 2008;12(11):1039–48.

    Article  Google Scholar 

  40. Moteghaed NY, Maghooli K, Pirhadi S, Garshasbi M. Biomarker discovery based on hybrid optimization algorithm and artificial neural networks on microarray data for cancer classification (in Eng). J Med Signals Sens. 2015;5(2):88–96.

    Article  PubMed  PubMed Central  Google Scholar 

  41. Moosa JM, Shakur R, Kaykobad M, Rahman MS. Gene selection for cancer classification with the help of bees (in Eng). BMC Med Genom. 2016;9(Suppl 2):47–47.

    Article  Google Scholar 

  42. Li B, Li J, Tang K, Yao X. Many-objective evolutionary algorithms. ACM Comput Surv. 2015;48:1–35.

    Article  Google Scholar 

  43. Yu PL. Cone convexity, cone extreme points, and nondominated solutions in decision problems with multiobjectives. J Optim Theory Appl. 1974;14(3):319–77.

    Article  Google Scholar 

  44. Theodoridis S, Koutroumbas K. Chapter 5—Feature selection. In: Theodoridis S, Koutroumbas K, editors. Pattern recognition. 4th ed. Boston: Academic Press; 2009. p. 261–322.

    Chapter  Google Scholar 

  45. Blondel V, Guillaume J-L, Lambiotte R, Lefebvre E. Fast unfolding of communities in large networks. J Stat Mech Theory Exp. 2008;2008:P10008.

    Article  Google Scholar 

  46. De Meo P, Ferrara E, Fiumara G, Ricciardello A. A novel measure of edge centrality in social networks. Knowl Based Syst. 2012;30:136–50.

    Article  Google Scholar 

  47. Qi X, Fuller E, Wu Q, Wu Y, Zhang C-Q. Laplacian centrality: a new centrality measure for weighted networks. Inf Sci. 2012;194:240–53.

    Article  Google Scholar 

  48. Fernández-Martínez JL. A brief historical review of Particle Swarm Optimization (PSO). J Bioinform Intell Control. 2012;1:3–16.

    Article  Google Scholar 

  49. M. Hall, E. Frank, G. Holmes, B. Pfahringer, P. Reutemann, I. Witten. The WEKA data mining software. http://www.cs.waikato.ac.nz/ml/weka.

  50. Moraglio A, Chio C, Togelius J, Poli R. Geometric particle swarm optimization. J Artif Evol Applicat. 2008;2008:1.

    Article  Google Scholar 

  51. Ghosh M, Begum S, Sarkar R, Chakraborty D, Maulik U. Recursive Memetic Algorithm for gene selection in microarray data. Expert Syst Appl. 2019;116:172–85.

    Article  Google Scholar 

  52. Prasad Y, Biswas KK, Hanmandlu M. A recursive PSO scheme for gene selection in microarray data. Appl Soft Comput. 2018;71:213–25.

    Article  Google Scholar 

  53. Chen K-H, Wang K-J, Wang K-M, Angelia M-A. Applying particle swarm optimization-based decision tree classifier for cancer classification on gene expression data. Appl Soft Comput. 2014;24:773–80.

    Article  Google Scholar 

  54. Friedman M. A comparison of alternative tests of significance for the problem of m rankings. Ann Math Stat. 1940;11(1):86–92.

    Article  Google Scholar 

  55. Nie NH, Hull CH, Jenkins JG, Steinbrenner K, Bent DH. Statistical package for the social sciences. New York: McGraw Hill; 1975.

    Google Scholar 

  56. Liu H, Yu L. Toward integrating feature selection algorithms for classification and clustering. IEEE Trans Knowl Data Eng. 2005;17(4):491–502.

    Article  Google Scholar 

  57. Chandrashekar G, Sahin F. A survey on feature selection methods. Comput Electr Eng. 2014;40(1):16–28.

    Article  Google Scholar 

  58. De-Stefano FFC, Marrocco C, Scotto di Freca A. A GA-based feature selection approach with an application to handwritten character recognition. Pattern Recognit Lett. 2014;35:130–41.

    Article  Google Scholar 

Download references

Acknowledgements

Not applicable.

Funding

The authors received no financial support for the research, authorship, and publication of this article.

Author information

Authors and Affiliations

Authors

Contributions

SA designed the proposed algorithm, AA validated the methodological approach and the research equations. SA and AA wrote literature review by reading over 30 articles and synthesizing them. SA wrote the first draft of the manuscript. AA improved the manuscript in English. All authors participated in the final review of the manuscript, corrected and approved the manuscript for submission. All authors read and approved the final manuscript.

Corresponding author

Correspondence to Saeid Azadifar.

Ethics declarations

Ethics approval and consent to participate

Not applicable.

Consent for publication

Not applicable.

Competing interests

The authors declare that they have no competing interests.

Additional information

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/. The Creative Commons Public Domain Dedication waiver (http://creativecommons.org/publicdomain/zero/1.0/) applies to the data made available in this article, unless otherwise stated in a credit line to the data.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Azadifar, S., Ahmadi, A. A graph-based gene selection method for medical diagnosis problems using a many-objective PSO algorithm. BMC Med Inform Decis Mak 21, 333 (2021). https://doi.org/10.1186/s12911-021-01696-3

Download citation

  • Received:

  • Accepted:

  • Published:

  • DOI: https://doi.org/10.1186/s12911-021-01696-3

Keywords