Skip to main content

Multi-task feature selection in microarray data by binary integer programming


A major challenge in microarray classification is that the number of features is typically orders of magnitude larger than the number of examples. In this paper, we propose a novel feature filter algorithm to select the feature subset with maximal discriminative power and minimal redundancy by solving a quadratic objective function with binary integer constraints. To improve the computational efficiency, the binary integer constraints are relaxed and a low-rank approximation to the quadratic term is applied. The proposed feature selection algorithm was extended to solve multi-task microarray classification problems. We compared the single-task version of the proposed feature selection algorithm with 9 existing feature selection methods on 4 benchmark microarray data sets. The empirical results show that the proposed method achieved the most accurate predictions overall. We also evaluated the multi-task version of the proposed algorithm on 8 multi-task microarray datasets. The multi-task feature selection algorithm resulted in significantly higher accuracy than when using the single-task feature selection methods.


Microarray technology has the ability to simultaneously measure expression levels of thousands of genes for a given biological sample, which is classified into one of the several categories (e.g., cancer vs. control tissues). Each sample is represented by a feature vector of gene expressions obtained from a microarray. Using a set of microarray samples with known class labels, the goal is to learn a classifier able to classify a new tissue sample based on its microarray measurements. A typical microarray classification data set contains a limited number of labeled examples, ranging from only a few to several hundred. Building a predictive model from such small-sample high-dimensional data is a challenging problem that has received a significant attention in machine learning and bioinformatics communities. To reduce the risk of over-fitting, a typical strategy is to select a small number of features (i.e., genes) before learning a classification model. As such, feature selection [1, 2] becomes an essential technique in microarray classification.

There are several reasons for feature selection in microarray data, in addition to improving the classifier's generalization ability. First, the selected genes might be of interest to domain scientists interested in identifying disease biomarkers. Second, building a classifier from a small number of features could result in an easily interpretable model that could give important clues to biologists. Depending on how the feature selection process is combined with model learning process, feature selection techniques can be organized into three categories. (1) Filter methods [3] are independent of the learning algorithm. (2) Wrapper methods [4] are coupled with the learning algorithm using heuristics such as forward selection and backward elimination. (3) Embedded methods [5, 6] integrate feature selection as a part of the classifier training. Both the wrapper and the embedded methods effectively introduce hyper-parameters that require computationally costly nested cross-validation and increase likelihood of over-fitting. Feature filter methods are very popular because they are typically conceptually simple, computationally efficient, and robust to over-fitting. These properties also explain why the filter methods are more widely used than the other two approaches in microarray data classification.

Traditional filter methods rank the features based on their correlation with the class label and then select the top ranked features. The correlation can be measured by statistic tests (e.g., t-test) or by information-theoretic criteria such as mutual information. The filter methods easily scale up to high dimensional data and can be used in conjunction with any supervised learning algorithm. However, because the traditional filter methods access each feature independently, highly correlated features tend to have similar rankings and tend to be selected jointly. Using redundant features could result in low classification accuracy. As a result, one common improvement for filter methods is to reduce redundancy between selected features. For example, minimal-redundancy-maximal-relevance (mRMR) proposed by [7] selects the feature set with both maximal relevance to the target class and minimal redundancy among the selected feature set. Because of the high computational cost of considering all possible feature sets, the mRMR algorithm selects features greedily, minimizing their redundancy with features chosen in previous steps and maximizing their relevance to the target class.

A common critique of popular feature selection filters is that they are typically based on relatively simple heuristics. To address this concern, recent research resulted in more principled formulation of feature filters. For example, algorithms proposed in [8] and [9] attempt to select the feature subset with maximal relevance and minimal redundancy by solving a constrained quadratic optimization problem (QP). The objective used by [8] is a combination of a quadratic term and a linear term. The redundancy between feature pairs is measured by the quadratic term and the relevance between features and class label is measured by the linear term. The features are ranked based on a weight vector obtained by solving a QP problem. The main limitation of this method is that the relevance between a feature and the class label is measured by either Pearson correlation or mutual information. However, Pearson correlation assumes normal distribution of the measurements, which might not be appropriate to measure correlation between numerical features and binary target. The mutual information requires using discrete variables and is sensitive to discretization. The objective used by [9] contains only one quadratic term. This quadratic term consists of two parts: one measures feature relevance using mutual information between features and the class label, and another measures feature redundancy using mutual information between each feature pair. However, the square matrix in the proposed quadratic term is not positive semi-definite. Thus, the resulting optimization problem is not convex and could result in poor local optima.

In this paper, we propose a novel feature filter method to find the feature subset which maximizes the inter-class separability and intra-class tightness, and minimizes the pairwise correlations between selected features. We formulate the problem as a quadratic programming with binary integer constraints. For high dimensional microarray data, to solve the proposed quadratic programming problem with binary integer constraints requires high time and space cost. Therefore, we relax binary integer constraints and apply the low rank approximation to the quadratic term in the objective function. The resulting objective function can be efficiently solved to obtain a small subset of features with maximal relevance and minimal redundancy.

In many real-life microarray classification problems, the size of the given microarray dataset is particularly small (e.g., we might have less than 10 labeled high-dimensional examples). In this case, even the most carefully designed feature selection algorithms are bound to underperform. Probably the only remedy is to borrow strength from external microarray datasets. Recent research [10, 11] illustrates that multi-task feature selection algorithms can improve the classification accuracy. The multi-task feature selection algorithms select the informative features jointly across many different microarray classification data sets. Following this observation, we extend our feature selection algorithm to the multi-task microarray classification setup.

The contributions of this paper can be summarized as follows.(1) We propose a novel gene filter method which can obtain a feature subset with maximal discriminative power and minimal redundancy; (2) The globally optimal solution can be found efficiently by relaxing the integer constraints and using a low-rank approximation technique; (3) We extend our feature selection method to multi-task classification setting; (4) The experimental results show our algorithms achieve higher accuracy than the existing filter feature selection methods, both in single-task learning and multi-task settings.

Results and discussion

We compared our proposed feature algorithm with 9 representative feature selection filters. The first 6 are standard feature selection filters: Pearson Correlation (PC), ChiSquare [3], GINI, Infogain, Kruskal-Wallis test and Relief [12]. They rank the features based on different criteria that measure correlation between each feature and class label. The remaining 3 are the state-of-the-art feature selection methods which are able to remove redundant features: mRMR [7], QPFS [8] and SASMIF [9]. The feature similarity for both QPFS and our algorithm was measured by Pearson correlation. For fair comparison, for the SASMIF method we used top m ranked features. To balance the effect of feature relevance and feature redundancy, the parameter λ in (9) was set to m 2 M i C i i , j Q i j . The low-rank parameter k was set to 0.1 · M, as suggested in [13]. Our algorithm is denoted as ST-BIP for single task version and MT-BIP for multi-task version.

Given the selected features, we used LIBLINEAR [14] to train the linear SVM model. The linear SVM model was chosen because previous studies [5] showed SVM classifier could be very accurate on microarray data. The regularization parameter C of LIBLINEAR was chosen among {10-3, 10-4, …, 103}. For the experiments in the single-task scenario, we used the nested 5 cross validation to select the optimal regularization parameter. For experiments in multi-task learning scenario, it was too time consuming to use the nested cross-validation to select the regularization parameter. Thus, we simply fixed the regularization parameter to 1 in the multi-task experiments.

Single task feature selection

In this section, we evaluate our proposed feature selection algorithm for single-task learning using four benchmark microarray gene expression cancer datasets: (1) Colon dataset [15] containing 62 samples, 40 tumor and 22 normal samples; (2) Lung dataset [16] containing 86 samples coming from 24 patients that died and 62 that survived; (3) Diffuse B-cell Lymphoma (DLBCL) dataset [17] containing 77 samples, 58 coming from DLBCL patients and 19 from Bcell lymphoma patients. (4) Myeloma dataset [18] containing 173 samples, 137 coming from patients with bone lytic lesions and 36 from control patients. We summarize the characteristics of these datasets in Table 1.

Table 1 Summary of the Microarray Datasets

For each microarray dataset, we randomly selected 20 positive and 20 negative examples (except for choosing 15 positive and 15 negative in DLBCL dataset) as the training set and the rest as the test set. Due to the class imbalance in test sets, we used AUC, the area under the Receiver Operating Characteristic (ROC) curve, to evaluate the performance. The average AUC based on 10 repetitions of experiments on different random splits to training and test set are reported in Table 2. We Compared the AUC accuracy of different feature selection algorithms for m = 20, 50, 100, 200, 1000. For each dataset, the best AUC score among all methods was emphasized in bold. As shown in Table 2, our proposed method achieved the highest accuracy on Colon and DLBCL datasets. On the Myeloma dataset, it had the highest accuracy when m = 100 and 1000 and had the second highest accuracy when m = 20, 50 and 200. On the Lung dataset, our algorithm was ranked in the upper half of the competing algorithms. The last column in Table 2 shows the average AUC score across four different datasets. Our method achieved the highest average AUC scores. The next two successful feature selection algorithms are Relief and QSFS. The mRMR had somewhat lower accuracy, comparable to simple filters such as PC, ChiSquare, GINI and InfoGain. SASMIF was considerably less accurate, while KW was the least successful.

Table 2 Average AUC of 10 different feature selection algorithms on 4 different microarray datasets

Multi-task feature selection

In this section, we evaluate our proposed feature selection algorithm for multi-task learning. We used 8 cancer related binary microarray classification datasets published in [19]. The data are summarized in Table 3. As shown in Table 3, the size of the 8 microarray datasets was very small. The single-task feature selection algorithms are not expected to perform well because there might be insufficient information even when simple feature selection filters are used. In contrast, our multi-task feature selection algorithm is expected to improve the accuracy by borrowing strength across multiple microarray datasets.

Table 3 Multi-Task Microarray Datasets(cancer:normal case)

For each microarray data set, we randomly selected N+ = 2, 3, 4, 5 positive and the same number of negative examples as the training data and used the rest as the test data. We show the results for m = 100 in this section. The average AUC across these 8 microarray datasets is shown in Figure 1. The results clearly show the multi-task version of our proposed algorithm was the most successful algorithm overall.

Figure 1

Average AUC score of different feature selection algorithms across different train sizes.

To gain a deeper understanding about the reason why the multi-task feature selection algorithm obtained better overall accuracy than single-task feature selection algorithms, we show the AUC score of each individual microarray dataset based on N + =3 in Table 4. We can see that the single task version of our feature selection algorithm had the highest overall accuracy among other single-task benchmarks, a result consistent with Table 2. The multi-task version of our algorithm has higher AUC than its single task version on 4 datasets and its average AUC is about 1.5% higher. In 4 cases, (e.g. Colon, Lung, Pancreas, Renal datasets) we can also observe the negative transfer, where the accuracy drops. How to prevent negative transfer in multi-task feature selection would be another interest research topic for our future research.

Table 4 Average AUC of 11 different feature selection algorithms on 8 different microarray datasets

Gene-annotation enrichment analysis for multi-task microarray datasets

The multi-task experimental results show that accuracies obtained by MT-BIP are better than other single task feature filters overall. So we would like to perform function annotation of the MP selected genes. In MT-BIP filter, only one selected gene list is obtained for all 8 different types of cancers. Given this gene list, the top 10 enriched GO terms were obtained using DAVID Bioinformatics Resources [20]. The top 10 enriched GO terms based on MT-BIP selected gene list is shown in Table 5. In this table, the hits means the number of genes that are found in the selected gene list associating with the specific GO term. The p-value was obtained by Fisher Exact test which is used to measure the gene-enrichment in annotation terms. After we got the enriched GO terms, we used the Comparative Toxicogenomics Database (CTD) [21] to check whether there is an association between the GO term and the cancer type. The last column in Table 5 shows the disease association for each GO term. The datasets are ordered as Bladder (B), Breast (B), Colon (C), Lung (L), Pancreas (P), Prostate (P), Renal (R) and Uterus (U). If a GO term is associated with the given type of cancer, we write down the cancer name. Otherwise, we put the symbol # in that position. We could see that the enriched GO terms based MT-BIP tends to associate many different types of cancer. As shown in Table 5, GO:0005856 (cytoskeleton), GO:0005886 (plasma membrane) and GO:0032403 (protein complex binding) were associated with 7 different cancers. GO:0030054 (cell junction) and GO:0015629 (actin cytoskeleton) are associated with 6 different cancers.

Table 5 Top 10 enriched GO terms based on 100 MT-BIP selected genes


We proposed a novel feature filter method to select a feature subset with discriminative power and minimal redundancy. The proposed feature selection method is based on quadratic optimization problem with binary integer constraints. It can be solved efficiently by relaxing the binary integer constrains and applying a low-rank approximation to the quadratic term in the objective. Furthermore, we extend our feature selection algorithm to multi-task classification problems. The empirical results on a number of microarray datasets show that in the single task scenario the proposed algorithm results in higher accuracy than the existing feature selection methods. The results also suggest that our multitask feature selection algorithm can further improve the microarray classification performance.


Feature selection by binary integer programming

Let us denote the training dataset as D = (x i , y i ), i = 1, …, N , where x i is an M dimensional feature vector for the i-th example and y i is its class label. N is the number of training examples. Our objective is to select a feature subset that is strongly predictive of class label and has low redundancy. We introduce a binary vector w = [ w 1 , w 2 , . . . , w M ] T to indicate which features are selected:

w j = 1 i f f e a t u r e j i s s e l e c t e d 0 i f f e a t u r e j i s n o t s e l e c t e d

So, the new feature vector for the i-th example after feature selection can be represented as g i = x i w, where the symbol denotes the pairwise product. Therefore, g ij = x ij , for w j = 1 and g ij = 0 for w j = 0. Alternatively, g i can be represented as g i = W x i , where W is a diagonal matrix and its diagonal is the vector w.

Intuitively, we would like the examples with the same class to be close (intra-class tightness) and the examples from different classes to be far away (inter-class separability) in the spaces defined by selected features. The Euclidean distance between two examples x i and x j in the new feature space can be calculated as

d i j = g i - g j 2 = x i w - x j w 2 = W x i - W x j 2

The inter-class separability of the data can be measured by a sum of the pairwise distances between examples with different class labels

y i y j x i w - x j w 2 .

The intra-class tightness of the data can be measured by a sum of the pair-wise distances between examples with the same class label

y i = y j x i w - x j w 2 .

Therefore, the problem of selecting a feature subset to maximize the intra-class tightness and inter-class separability can be formulated as

min w y i = y j x i w - x j w 2 - y i y j x i w - x j w 2 .

Objective (5) can be rewritten as

min w i = 1 N j = 1 N x i w - x j w 2 A i j ,

where matrix A is defined as:

A i j = 1 i f y i = y j - 1 i f y i y j

In addition to the objective (5) or (6), in order to improve the diversity of selected features, we would like to select a feature subset with minimal redundancy. A feature is defined to be redundant if there is another feature highly correlated with it. Let us denote Q as a symmetric positive semidefinite matrix with size M × M, whose element Q ij represents the similarity between feature i and feature j. Since the measurements of each feature across different samples are normal distributed, it is reasonable to use Pearson Correlation to measure the similarity between two features here. Also, the similarity matrix Q is positive semi-definite when Pearson Correlation is used. Then, we define a redundancy among the selected set of features represented by vector w as their average pair-wise similarity wTQ w /m2, where m is the number of selected features. Our objective is to minimize the redundancy defined in such way.

The first contribution of this paper is to formulate the feature selection task as a new quadratic programming problem subject to binary integer and linear constraints as follows,

min w i = 1 N j = 1 N x i w - x j w 2 A i j + λ 1 m 2 w T Q w s . t . w i { 0 , 1 } i i = 1 M w i = m .

The first term in (8), which is a linear term as shown in the following Proposition 1, tries to maximize the interclass separability and intra-class tightness of the data. It describes the discriminative power of the selected feature subset. The second quadratic term is the average pair-wise similarity score between the selected features, which results in reduction of feature redundancy. Parameter λ is introduced to control the tradeoff between feature relevance and feature redundancy. Since Q is a positive semi-definite matrix, the proposed objective function is convex. The first constraint ensures that the resulting vector w is binary, while the second constraint ensures that exactly m features are selected. The following proposition establishes that the first term in the objective (8) is linear.

Proposition 1. The first term of the objective function (8) can be written as a linear term cT w , where c is vector of size M with elements c i = (XTLX) ii , L is the Laplacian matrix of A, defined as L = DA. D is a diagonal degree matrix such that D i i = j A i j . The X is the N × M feature matrix. Each row in X corresponds to one example. (XTLX) ii denotes the i-th element in the diagonal of the matrix XTLX.

Proof. Let us denote W as a diagonal matrix where W ii = w i . Then,

i = 1 N j = 1 N x i w - x j w 2 A i j = i = 1 N j = 1 N W x i - W x j 2 A i j = t r a c e ( W T X T L X W ) = t r a c e ( X T L X W W T )

because w i {0, 1}, WWT = W. Therefore, trace ( X T L X W W T ) = i = 1 M ( X T L X ) i i W i i = c T w , where c i = (XT LX) ii

Based on Proposition 1, objective (8) can be rewritten as the following constrained quadratic optimization problem,

min w c T w + λ 1 m 2 w T Q w s . t . w i { 0 , 1 } i i = 1 M w i = m .

There are two practical obstacles in solving (9): (1) Binary constraint of variable w, and (2) feature similarity matrix Q is with size M × M, which implies high computational cost for high dimensional data. In the next two sections, we will first relax the binary constraint, and then we will apply a low-rank approximation to Q. The resulting constrained optimization problem can be solved very efficiently, with linear time with respect to the number of features M.

Problem Relaxation. Due to the binary constraint on the indictor vector w, it is difficult to solve (9) [9]. To resolve this, we first relax the binary constraint on w by allowing its elements w i to be within the range [0, m]. Then, (9) could be approximated by

min w c T w + λ 1 m 2 w T Q w s . t . w i 0 i i = 1 M w i = m .

Now, (10) becomes a standard Quadratic Programming (QP) problem. The optimal solution can be obtained by a general QP solver (e.g., MOSEK [22]).

Low-rank Approximation. The matrix Q in (10) is of size M × M . So, it results in high time and space cost if we work with high dimensional microarray data. Therefore, we would like to avoid the computational bottleneck by using low-rank approximation techniques.

The matrix Q in (10) is symmetric positive semidefinite. So, it can be decomposed as Q = UΛUT, where U is a matrix of eigenvectors and Λ is a diagonal matrix with corresponding eigenvalues of Q. By setting α= Λ 1 2 U T w , it follows that w =U Λ - 1 2 α. Therefore, problem (10) can be rewritten as

min α c T U Λ - 1 2 α + λ 1 m 2 α T α s . t . U Λ - 1 2 α 0 1 U Λ - 1 2 α = m .

Typically, the rank of Q (let us denote it as k) is much smaller than M, kM. Therefore, we can replace the full eigenvector and eigenvalue matrices U and Λ by the top k eigenvectors and eigenvalues, resulting in an M × k matrix U k and a k × k diagonal matrix Λ k , without losing much information. Therefore, (11) is reformulated as

min α c T U k Λ k - 1 2 α + λ 1 m 2 α T α s . t . U k Λ k - 1 2 α 0 1 U k Λ k - 1 2 α = m .

Since α is a vector with length k, kM. the QP (11) is reduced to a new QP in a k-dimensional space with M + 1 constraints. Once the solution α of (12) is obtained, the variable w in original space can be approximated by w = U k Λ k - 1 2 α .

Decomposing matrix Q requires O(M3) time, which is expensive in microarray data where M is large. Next we will show how to efficiently compute the top k eigenvectors and eigenvalues using Nystrom approximation technique [23]. Nystrom method approximates a M × M symmetric, positive semi-definite matrix Q by

Q = E M k W k k - 1 E M k T

where E Mk denotes the sub-matrix of Q created by selecting k of its columns, and W kk is a sub-matrix that corresponds to the intersection of the selected columns and rows. Sampling schemes in Nystrom method include random sampling [23], probabilistic sampling [24], and k-means based sampling [13]. We chose the k-means sampling in our experiments because [13] showed that it produces very good low-rank approximations at a relatively low cost. Given (13), we can easily obtain the low rank approximation of Q as

Q = G G T w h e r e G = E M k W k k - 1 2 .

As shown in the following Proposition 2, the top k eigenvectors and eigenvalues can be computed in O(Mk2) time using Nystrom method, which is much more efficient than doing eigen-decomposition of Q, which requires O(M3) time.

Proposition 2. The top k eigenvectors U k and the corresponding eigenvectors Λ k of Q = G G T can be approximated as Λ k = Λ G and U k =G U G Λ G - 1 2 , where U G and Λ G are obtained by the eigen-decomposition of k × k matrix G T G= U G Λ G U G T

Proof. First, we observe that U k contains orthonormal columns.

U k T U k = Λ G - 1 2 U G T G T G U G Λ G - 1 2 = Λ G - 1 2 U G T U G Λ G U G T U G Λ G - 1 2 = I .

Next, we observe that

U k Λ k U k T = G U G Λ G - 1 2 Λ G Λ G - 1 2 U G T G T = G G T = Q

Our proposed feature selection algorithm is summarized in Algorithm 1. In the Algorithm 1, steps 1 to 5 require O(Mk2 + k3) time. QP in step 6 with k variables has a polynomial time complexity with respect to k. Step 7 requires O(Mk) time. Therefore, overall, the proposed feature selection algorithm is very efficient and it has linear time complexity with the number of features M.

Algorithm 1 Single-Task Binary Integer Program Feature Selection

Input: training data X, their labels y, regularized parameter λ, number of features m, low-rank parameter k.

Output: m selected features

  1. 1.

    Apply Proposition 1 to compute the vector c

  2. 2.

    Use k-means to select k landmark features for low-rank approximation of Q

  3. 3.

    Compute E Mk and W kk in (13)

  4. 4.

    Obtain low-rank approximation of Q by (14)

  5. 5.

    Apply Proposition 2 to compute the top k eigenvalue Λ k and eigenvector U k of Q

  6. 6.

    Obtain α by solving the lower dimensional QP problem(12).

  7. 7.

    Obtain w in original feature space as w = U k Λ - 1 2 α

  8. 8.

    Rank the features according to the weight vector w and select the top m features

Multi-task feature selection by binary integer programming

Multi-task learning algorithms have been shown to be able to achieve significantly higher accuracy than single task learning algorithms both empirically [11] and theoretically [25]. Motivated by these promising results, in this section, we extend our feature selection algorithm to the multi-task setting. The objective is to select features which are discriminative and non-redundant over multiple microarray datasets.

Let us suppose there are K different but similar classification tasks, and denote the training data of the t-th task as D t = { ( x i t , y i t ) , i = 1 , . . . , N t } , where N t is the number of training examples of the t-th task. [10, 11] proposed multi-task feature selection algorithms that use 1,2 norm to regularize the linear model coefficients β across K different classification tasks. The 1,2 norm regularizer over all βs across K classification tasks could be expressed as j = 1 M ( t = 1 K β t j 2 ) , where β t j is the coefficient of the j-th feature in the t-th task. Due to the 1 norm on the 2 norm of group of coefficients of each feature across K tasks, the 1,2 norm regularizer selects the same feature subset across K tasks. However, the 1,2 norm regularized problem is challenging to solve because the non-smoothness of the 1,2 norm. In this section, we would like to show our proposed feature selection can be easily extended to multi-task learning version. The resulting objective optimization problem have the same form as objective (9), which can be solved efficiently as shown in previous section.

Let us denote w t as the binary indicator defined in (1) to represent the selected feature subset of the t-th classification task. If we do not consider the relatedness between these K classification task, individual w t could be obtained by applying Algorithm 1 to different classification tasks. Based on the conclusion given by [10, 11], it would be beneficial to select the same feature subset across K related classification task. In our case, this is can be achieved by setting w t = w t. Therefore, the same feature across K tasks, defined by vector w, can be obtained by solving the following optimization problem,

min w j = 1 K c j T w + λ 1 m 2 w T j = 1 K Q j w s . t . w i { 0 , 1 } i i = 1 M w i = m .

where c j and Q j are the linear and quadratic terms of the QP corresponding to the j-th task. The details about how to compute the c j and Q j are explained in the previous section. The technique of relaxing binary integer constraints and applying low-rank approximation to Q introduced in the previous section can be used to solve (15). The extended multi-task feature selection algorithm is also a feature filter. It can be used in conjunction with any supervised learning algorithm.


  1. 1.

    Guyon I, Elisseeff A: An Introduction to Variable and Feature Selection. Journal of Machine Learning Research. 2003, 1157-1182.

    Google Scholar 

  2. 2.

    Saeys Y, Inza I, Larranãga P: A review of feature selection techniques in bioinformatics. Bioinformatics. 2007, 23 (19): 2507-2517. 10.1093/bioinformatics/btm344.

    CAS  Article  PubMed  Google Scholar 

  3. 3.

    Liu H, Setiono R: A probabilistic approach to feature selection - a filter solution. Proceedings of the Thirteeth th International Conference on Machine Learning. 1996, 319-327.

    Google Scholar 

  4. 4.

    Kohavi R, John G: Wrappers for Feature Subset Selection. Artificial Intelligence. 1997, 97: 273-324. 10.1016/S0004-3702(97)00043-X.

    Article  Google Scholar 

  5. 5.

    Guyon I, Weston J, Barnhill S, Vapnik V: Gene Selection for Cancer Classification using Support Vector Machines. Machine Learning. 2002, 46: 389-422. 10.1023/A:1012487302797.

    Article  Google Scholar 

  6. 6.

    Tibshirani RJ: Regression Shrinkage and Selection via the Lasso. Journal of the Royal Statistical Society, Series B. 1996, 58: 267-288.

    Google Scholar 

  7. 7.

    Peng H, Long F, Ding CHQ: Feature Selection Based on Mutual Information: Criteria of Max-Dependency, Max-Relevance, and Min-Redundancy. IEEE Transactions on Pattern Analysis and Machine Intelligence. 2005, 27 (8): 1226-1238.

    Article  PubMed  Google Scholar 

  8. 8.

    Lujan IR, Huerta R, Elkan C, Cruz CS: Quadratic Programming Feature Selection. Journal of Machine Learning Research. 2010, 11: 1491-1516.

    Google Scholar 

  9. 9.

    Liu S, Liu H, Latecki LJ, Yan S, Xu C, Lu H: Size Adaptive Selection of Most Informative Features. Proceedings of the Twenty-Fifth AAAI Conference on Artificial Intelligence. 2011

    Google Scholar 

  10. 10.

    Argyriou A, Evgeniou T, Pontil M: Convex multi-task feature learning. Machine Learning. 2008, 73 (3): 243-272. 10.1007/s10994-007-5040-8.

    Article  Google Scholar 

  11. 11.

    Obozinski G, Taskar B, Jordan MI: Joint covariate selection and joint subspace selection for multiple classification problems. Statistics and Computing. 2010, 20 (2): 231-252. 10.1007/s11222-008-9111-x.

    Article  Google Scholar 

  12. 12.

    Kira K, Rendell LA: A Practical Approach to Feature Selection. Proceedings of the Ninth International Conference on Machine Learning. 1992, 249-256.

    Google Scholar 

  13. 13.

    Zhang K, Kwok JT, Parvin B: Prototype vector machine for large scale semi-supervised learning. Proceedings of the Twenty-sixth International Conference on Machine Learning. 2009, 1233-1240.

    Google Scholar 

  14. 14.

    Fan RE, Chang KW, Hsieh CJ, Wang XR, Lin CJ: LIBLINEAR: A Library for Large Linear Classification. Journal of Machine Learning Research. 2008, 9: 1871-1874.

    Google Scholar 

  15. 15.

    Alon U, Barkai N, Notterman DA, Gish K, Ybarra S, Mack D, Levine AJ: Broad patterns of gene expression revealed by clustering analysis of tumor and normal colon tissues probed by oligonucleotide arrays. The Proceedings of the National Academy of Sciences USA. 1999, 6745-6750.

    Google Scholar 

  16. 16.

    Beer DG, Kardia SL, Huang CC, Giordano TJ, Levin AM, Misek DE, Lin L, Chen G, Gharib TG, Thomas DG, Lizyness ML, Kuick R, Hayasaka S, Taylor JM, Iannettoni MD, Orringer MB, Hanash S: Gene-expression profiles predict survival of patients with lung adenocarcinoma. Nature Medicine. 2002, 816-824.

    Google Scholar 

  17. 17.

    Shipp MA, Ross KN, Tamayo P, Weng AP, Kutok JL, Aguiar RC, Gaasenbeek M, Angelo M, Reich M, Pinkus GS, Ray TS, Koval MA, Last KW, Norton A, Lister TA, Mesirov J, Neuberg DS, Lander ES, Aster JC, Golub TR: Diffuse large B-cell lymphoma outcome prediction by gene-expression profiling and supervised machine learning. Nature Medicine. 2002, 68-74.

    Google Scholar 

  18. 18.

    Tian E, Zhan F, Walker R, Rasmussena E, Ma Y, Barlogie B, Shaughnessy J: The role of the Wnt-signaling antagonist DKK1 in the development of osteolytic lesions in multiple myeloma. New England Journal of Medicine. 2003, 2483-2494.

    Google Scholar 

  19. 19.

    Ramaswamy S, Tamayo P, Rifkin R, Mukherjee S, Yeang CH, Angelo M, Ladd C, Reich M, Latulippe E, Mesirov JP, Poggio T, Gerald W, Loda M, Lander ES, Golub TR: Multiclass cancer diagnosis using tumor gene expression signatures. Proceedings of the National Academy of Sciences USA. 2001, 15149-54.

    Google Scholar 

  20. 20.

    Da Wei Huang BTS, Lempicki RA, et al: Systematic and integrative analysis of large gene lists using DAVID bioinformatics resources. Nature protocols. 2008, 4: 44-57. 10.1038/nprot.2008.211.

    Article  Google Scholar 

  21. 21.

    Davis AP, King BL, Mockus S, Murphy CG, Saraceni-Richards C, Rosenstein M, Wiegers T, Mattingly CJ: The comparative toxicogenomics database: update 2011. Nucleic acids research. 2011, 39 (suppl 1): D1067-D1072.

    PubMed Central  CAS  Article  PubMed  Google Scholar 

  22. 22.

    Andersen ED, Andersen KD: The MOSEK interior point optimizer for linear programming: an implementation of the homogeneous algorithm. High Performance Optimization. 2000, 197-232.

    Google Scholar 

  23. 23.

    Williams CKI, Seeger M: The Effect of the Input Density Distribution on Kernel-based Classifiers. Proceedings of the Seventeenth International Conference on Machine Learning. 2000, 1159-1166.

    Google Scholar 

  24. 24.

    Drineas P, Mahoney MW: On the Nyström Method for Approximating a Gram Matrix for Improved Kernel-Based Learning. Journal of Machine Learning Research. 2005, 6: 2153-2175.

    Google Scholar 

  25. 25.

    Ben-David S, Schuller R: Exploiting Task Relatedness for Multiple Task Learning. COLT: Proceedings of the Workshop on Computational Learning Theory. 2003

    Google Scholar 

Download references


This work was supported by the U.S. National Science Foundation Grant IIS-0546155.


Publication of this work was supported by the U.S. National Science Foundation Grant IIS-0546155.

This article has been published as part of BMC Proceedings Volume 7 Supplement 7, 2013: Proceedings of the Great Lakes Bioinformatics Conference 2013. The full contents of the supplement are available online at

Author information



Corresponding author

Correspondence to Slobodan Vucetic.

Additional information

Competing interests

The authors declare that they have no competing interests.

Authors' contributions

LL and SV conceived the study and developed the algorithm. LL wrote the first draft of the manuscript. Both authors participated in the preparation of the manuscript and approved the final version.

Rights and permissions

This article is published under license to BioMed Central Ltd. This is an Open Access article distributed under the terms of the Creative Commons Attribution License (, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited. The Creative Commons Public Domain Dedication waiver ( applies to the data made available in this article, unless otherwise stated.

Reprints and Permissions

About this article

Cite this article

Lan, L., Vucetic, S. Multi-task feature selection in microarray data by binary integer programming. BMC Proc 7, S5 (2013).

Download citation


  • Feature Selection
  • Feature Subset
  • Feature Selection Method
  • Feature Selection Algorithm
  • Quadratic Programming