Figures
Abstract
Advances in DNA microarray technologies have made gene expression profiles a significant candidate in identifying different types of cancers. Traditional learning-based cancer identification methods utilize labeled samples to train a classifier, but they are inconvenient for practical application because labels are quite expensive in the clinical cancer research community. This paper proposes a semi-supervised projective non-negative matrix factorization method (Semi-PNMF) to learn an effective classifier from both labeled and unlabeled samples, thus boosting subsequent cancer classification performance. In particular, Semi-PNMF jointly learns a non-negative subspace from concatenated labeled and unlabeled samples and indicates classes by the positions of the maximum entries of their coefficients. Because Semi-PNMF incorporates statistical information from the large volume of unlabeled samples in the learned subspace, it can learn more representative subspaces and boost classification performance. We developed a multiplicative update rule (MUR) to optimize Semi-PNMF and proved its convergence. The experimental results of cancer classification for two multiclass cancer gene expression profile datasets show that Semi-PNMF outperforms the representative methods.
Citation: Zhang X, Guan N, Jia Z, Qiu X, Luo Z (2015) Semi-Supervised Projective Non-Negative Matrix Factorization for Cancer Classification. PLoS ONE 10(9): e0138814. https://doi.org/10.1371/journal.pone.0138814
Editor: Ramin Homayouni, University of Memphis, UNITED STATES
Received: March 4, 2015; Accepted: September 3, 2015; Published: September 22, 2015
Copyright: © 2015 Zhang et al. This is an open access article distributed under the terms of the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited
Data Availability: All relevant data are within the paper and its Supporting Information files. All relevant data are available on the Zenodo repository, and the access URL is https://zenodo.org/record/21712.
Funding: This work was supported by Ministry of Education of the People’s Republic of China Grant Number 20134307110017 (www.moe.edu.cn), PI: Zhigang Luo and National Natural Science Foundation of China Grant Number 91024030/G03 (http://www.nsfc.gov.cn/), PI: Xiaogang Qiu.
Competing interests: The authors have declared that no competing interests exist.
Introduction
In cancer prognosis and treatment, it is crucial to identify different cancer types and subtypes. Traditional methods often rely on similar morphological appearances but easily induce different clinical courses and responses to therapy because of subjective interpretations and personal experience. This usually results in diagnostic confusion. Fortunately, the emergence of the DNA microarray technique removes this barrier in an objective and systematic manner and has showed great potential in outcome prediction of cancer types in genome-wide scales [1–11].
Numerous learning methods have been developed for cancer classification based on gene expression profiles [1–3]. For instance, Golub et al. [1] used a weighted voting scheme for the molecular classification of acute leukemia. Nguyen et al. [3] incorporated partial least squares (PLS) into the logistic discrimination and quadratic discriminant analysis for tumor classification. However, these methods are not convenient for practical applications because the labeled samples are quite expensive in the clinical cancer research community. To overcome this deficiency, Xu et al. [12] used the semi-supervised Ellipsoid ARTMAP (ssEAM) method for cancer classification. Shi et al. [13] utilized the semi-supervised method termed low density separation (LDS, [14]) to classify different types of cancers. Moreover, Maulik et al. [15] investigated the effectiveness of transductive SVM (TSVM, [16]) in cancer classification. Nevertheless, these algorithmic challenges involve the curse of dimensionality, which indicates that the overwhelming number of measures for gene expression levels contrast with the small number of samples.
This problem often calls for dimension reduction techniques. This paper focuses on non-negative matrix factorization (NMF, [17, 18]) because it is a flexible framework for conducting dimension reduction and performing classification and clustering tasks [19–26]. NMF decomposes a data matrix into the product of two non-negative factors. Due to its effectiveness, NMF and its variants have been applied to analyzing large-scale gene expression datasets [27–29], cancer classification [30, 31] and new class discovery [30]. Brunet et al. [31] originally adopted NMF to uncover molecular meta-patterns by clustering samples of leukemia, medulloblastoma and central nervous system tumors, and indicating that NMF outperforms both hierarchy clustering (HC) and self-organizing map (SOM). However, NMF does not explicitly guarantee the sparseness of the decomposition and violates the uniqueness property. Recent works [32] show that this often degrades the clustering performance. To address this issue, Li et al. [32] proposed local NMF (LNMF) to overcome this deficiency by imposing the sparse constraints over the decomposition. Hoyer et al. proposed sparse NMF (SNMF, [33]) to enforce sparseness in NMF by penalizing the number of non-zero entries of the coefficients rather than the sum of the entries. Furthermore, Gao et al. [34] utilized SNMF to identify the meta-patterns of various cancers for identifying different types of tumors.
Because the aforementioned methods follow regularization theory, they are jointly non-convex and are difficult to optimize. Unlike the above methods, Yuan et al. [35] developed the projective NMF (PNMF) to induce parts-based representation by implicitly imposing the orthogonal constraint over the basis. However, because these methods are unsupervised learning methods that do not take into account labels, their performance in cancer classification can be further improved. In this paper, we propose a semi-supervised projective NMF method (Semi-PNMF) that utilizes both labeled and unlabeled samples to boost classification performance. Particularly, Semi-PNMF learns a non-negative subspace from concatenated labeled and unlabeled samples and predicts classes by the index of the largest entries of their coefficients. Benefiting from the unlabeled data, Semi-PNMF can learn more representative subspaces, which are beneficial for classification tasks. We explored a multiplicative update rule (MUR) to solve Semi-PNMF and proved its convergence. The experimental results of cancer identification for multiclass cancer gene expression profile datasets including GCM [8] and Acute Leukemia [36] datasets show that Semi-PNMF outperforms the representative methods in terms of quantity.
Materials and Methods
Semi-supervised Projective Nonnegative Matrix Factorization
Projective non-negative matrix factorization (PNMF) learns a non-negative projection matrix to project high-dimensional data into the lower-dimensional subspace. Because it can learn parts-based representation, PNMF has been widely applied in pattern recognition [21, 26, 35, 37]. Here, we introduce the other representation form of PNMF that learns the lower-dimensional coefficients of samples to approximate the class indicator for clustering. This is based on the assumption that the basis lies in the subspace spanned by the original samples. Given the data matrix V = [v1,⋯,vn]T ∈ Rn × m, where n denotes the number of samples and m their dimensionality, PNMF learns the coefficients H ∈ Rn × r to represent original samples, i.e.,
(1)
where ∥•∥F denotes the matrix Frobenius norm and r the number of clusters.
As in objective (1), it is non-trivial to analyze the convergence in theory because Eq (1) contains a fourth-order term. To remove such a high order term, we first introduce an auxiliary variable, i.e., the cluster centroids, and the equality constraint into Eq (1). Thus, we can obtain
(2)
The objective is very similar to BPNMF [26], but we cannot directly apply the optimization algorithm of BPNMF to optimize it especially when additional constraints such as the sparseness constraint and Laplacian regularization are imposed over the coefficients, as these constraints easily induce PNMF to produce the trivial solution. To avoid such a drawback, we propose a semi-supervised PNMF method (Semi-PNMF) by recasting Eq (2) as
(3)
where α ≥ 0 is a regularization constant and W denotes the non-negative cluster centroid. Model (3) significantly differs from BPNMF because Eq (3) favors the representative capacity of the cluster centroids, while BPNMF focuses on the orthogonality of the non-negative subspace. Thus, Eq (3) induces the sparse coefficients, while BPNMF produces the sparse basis.
According to Eq (3), we can incorporate the local coordinate constraint [38] to improve the representative power of the basis, meanwhile further inducing the sparse coefficients to be true classes. Thus, we recast Eq (3) as the following regularization form:
(4)
where β trades off the local coordinate regularization and Hij denotes the i-the row and j-th column element of coefficients H, Wj and Vi, signifying the i-th and j-th row vector of W and V, respectively.
To make full use of partial labeled samples, we propagate the labels of labeled samples to unlabeled ones by minimizing the distance between their coefficients and the corresponding class indicator. Particularly, we require the coefficients of labeled samples to be equivalent with the corresponding class indicator. Consider the first d examples labeled and the rest unlabeled; the data matrix V can be divided into two parts, i.e., . Then, we can obtain the objective function of Semi-PNMF as follows:
(5)
where Q denotes the partial label matrix wherein Qij = 1 if vi belongs to the j-th class; otherwise, Qij = 0. Both HU and nU denote the coefficients and number of the unlabeled samples, respectively.
Interestingly, Semi-PNMF has two distinct aspects. First, it replaces the learned coefficients of the labeled samples with the corresponding class indicator. The constraint is so strong that the learned basis completely biases the labeled samples. This might induce the trivial solution to the coefficients of the unlabeled samples. Second, Semi-PNMF completely ignores the representation contribution of the labeled samples. It is so unintelligible that the learned basis only favors the unlabeled samples. It appeared that both aspects contradict each other, but intrinsically, they mutually complement each other in our Semi-PNMF. In essence, the first aspect corresponds to supervised learning, which generates the reasonable solution yet does not ensure it is consistent with the underlying data distribution, while the second one considers data distribution but cannot yield the reasonable solution. Thus, the combination of both aspects can mutually complement each other. Semi-PNMF learns the shared basis by the labeled and unlabeled instances, meanwhile inducing similar instances to have a similar representation, i.e., the coefficients. Because we impose the restriction that coefficients of the labeled samples be their labels as well as the local coordinate constraint over the basis and coefficients, the unlabeled sample coefficients are implicitly as sparse as the label vectors. In this way, Semi-PNMF effectively propagates the labels of labeled samples to the unlabeled ones. Consequently, in cancer classification, it is reasonable that, for each unlabeled sample, we choose the index of the largest entry of its coefficient to predict the classes of this sample once objective (5) yields their coefficients. The above intuition can be further verified by the toy example given in Figs 1 and 2.
In Fig (a), both the square and circle markers signify the unlabeled and labeled samples, respectively, and three different colors stand for three different categories. Fig (b) shows that the unlabeled samples are marked as the ground-truth markers and colors. Figs (c) and (d) shows the coefficients and basis learned by Semi-PNMF, respectively. The index of maximum value of the coefficient for an unlabeled sample appears in red and indicates its class.
Optimization Algorithm
It is difficult to optimize Eq (5) because it is jointly non-convex with respect to both W and H. Fortunately, it is convex with respect to W and H, respectively. Thus, we can establish the following theorem:
Theorem 1: The objective function (5) is non-increasing under the following multiplicative update rules:
(6)
, and
(7)
where ⊗ denotes the element-wise product operator, FU = diag(sum(HU)), A = [a,⋯,a] wherein
, and B = [b,⋯,b], wherein b = diag(WWT).
Proof. According to Eq (5), we can obtain the objective with respect to W as follows:
(8)
where
denotes the diagonal matrix whose diagonal elements are the i-th row vector values of VU.
By Eq (8), we can define the auxiliary function of J(W) as
(9)
Obviously, objective (9) has
(10)
We can obtain the derivative of Eq (9) as follows:
(11)
Based on Eq (11), we have
(12)
By simple algebra, the formula (6) can be deduced from Eq (12). Likewise, we can obtain the auxiliary function of J(HU) as follows:
(13)
Thus, according to Eq (14), we also obtain the update rule (7) for HU.
Moreover, according to Eqs (10), (12) and (14), we have
(15)
Based on Eq (15), these update rules always guarantee that the objective function monotonically decreases. Thus, this completes the proof. ■
According to the above theorem, we summarize the multiplicative update rule (MUR) for Semi-PNMF in Algorithm 1.
Algorithm 1 MUR for Semi-PNMF
Input: Examples V ∈ Rm × n, penalty parameter α, partial label matrix Q.
Output: HU.
1: Randomly initialize W0 and , and l = 0.
2: repeat
3: Update Wl+1 according to Eq (6).
4: Calculate according to Eq (7).
5: l ← l+1.
6: until {Stopping criterion is satisfied.}
7: .
To reduce the time overhead, Algorithm 1 utilizes the objective relative error as the stopping criterion; in addition, set ɛ to 10−7 in our experiments. The main time cost of Algorithm 1 lies in line 3 and line 4. Their time complexities are O(r2n+mrn+r2m+rm) and O(mr(n − d)+r2m+rm+r2+r2(n − d)), respectively. Thus, the total time complexity of Algorithm 1 is O(r2 n+mrn+mr(n − d)+mrd+r2m+rm+r2+r2(n − d)).
Results
This section conducts a series of experiments on both synthetic and real-world datasets to verify the method proposed in this paper.
Synthetic Dataset
This section generates a small synthetic dataset to clarify the mechanism of Semi-PNMF. The synthetic dataset consists of three categories constructed by the following random samples:
and
where x ∈ R3, and each of its entry is sampled from the standard uniform distribution U(0,1). For each category, we randomly generated 10 samples, within which three samples were selected as labeled samples and the rest as unlabeled ones. Therefore, the synthetic dataset contains 30 samples in total. For clear illustration, three categories are marked as three different colors, and the labeled and unlabeled samples are distinguished by two shapes.
Fig 1(a) and 1(b) shows the ground truth and resultant labeled results of the unlabeled samples by Semi-PNMF, respectively, while Fig 1(c) and 1(d) displays the learned coefficients of the unlabeled samples and basis. In Fig 1(d), each row of the learned basis has different colors, implying that the basis stands for the centroids of different categories and owns the discriminative representation ability. According to Fig 1(c), each row of the learned coefficients is the lower-dimensional coefficient of the corresponding unlabeled sample. The larger the entry of the coefficient is, the darker its color is. As shown in Fig 1(c), the maximum entry of the coefficient largely exceeds the other entries. All maximum entries make the coefficients take up the diagonal form and imply the cluster memberships of all the samples. Thus, it is reasonable to select the index of the maximum entry of the coefficient as the classes of an unlabeled sample. This verifies our previous intuition. Since all samples shares the common basis, their coefficients become close to each other if they have the same labels. We impose the restriction that the coefficients of labeled samples be equivalent to their label vectors, and thus this also induces the coefficients of the unlabeled to be close to their label vectors. In this way, Semi-PNMF can propagate the labels of the labeled samples to the unlabeled ones. The propagation procedure is illustrated in Fig 2.
GCM Dataset
This experiment merely compares traditional semi-supervised learning methods including low density separation (LDS, [14]), transductive SVM (TSVM, [16]), constrained NMF (CNMF, [24]), soft-constrained NMF (SCNMF, [25]) and Semi-PNMF by separating different types of cancers on the GCM dataset. The GCM dataset [8] contains the expression profiles of 218 tumor samples representing 14 common human cancer classes. It is available on the public website: http://www.broadinstitute.org/cgi-bin/cancer/datasets.cgi, and can also be downloaded from the website: https://zenodo.org/record/21712. According to [8], we combine the training and testing set of this gene expression data into a dataset for cancer classification. Thus, the combined dataset contains 198 samples with 16,063 genes. Table 1 gives a brief description of this dataset. To remove very low noisy values and saturation effects of very high values, we bound the gene expression data into a specific box constraint ranging from 20 to 16,000 units and then exclude those genes whose ratios and absolute variations across samples are under 5 and 500, respectively. Consequently, the resultant expression profile dataset contains the 11,370 genes passing. We compare the effectiveness of Semi-PNMF with LDS, TSVM, CNMF and SCNMF under varying configurations. Both CNMF and SCNMF involve no parameter tuning. For Semi-PNMF, we set two parameters α = 2, and β = 0.0001, respectively. Because these representative methods enable convergence within 1,500 iteration rounds, we set the maximum number of loops to 1,500. For LDS and TSVM, we adopt the parameter settings provided in the source code to obtain the classification results.
We evaluate the cancer classification by the cross-validation over the whole dataset. This process selects one sample as the unlabeled sample and, meanwhile, learns the prediction model on all the samples for cancer diagnosis. For the unlabeled sample, we choose the index of the largest value of the resultant consensus matrix to predict the classes of this sample. As shown in Figs 3 to 7, the confusion matrix of the predicted results of Semi-PNMF, CNMF, SCNMF, LDS and TSVM are reported in detail. Each column denotes how many the unlabeled samples are assigned to each cancer, while each row signifies the number of the unlabeled samples affiliated to the real tumor type. Each color not only represents a specific cancer type but also highlights the correct prediction results, i.e., the diagonal elements of the confusion matrix.
Matrix delineates distribution of actual compared with predicted class membership for multiclass cancer prediction on the GCM dataset.
Matrix delineates distribution of actual compared with predicted class membership for multiclass cancer prediction on the GCM dataset.
Matrix delineates distribution of actual compared with predicted class membership for multiclass cancer prediction on the GCM dataset.
Matrix delineates distribution of actual compared with predicted class membership for multiclass cancer prediction on the GCM dataset.
Matrix delineates distribution of actual compared with predicted class membership for multiclass cancer prediction on the GCM dataset.
Figs 3 to 7 imply that Semi-PNMF can identify different tumor types more accurately than the representative methods. For example, when working with two labeled samples from each tumor type, Semi-PNMF achieves 70.71% classification accuracy and exceeds LDS, TSVM, SCNMF, and CNMF by 10.6%, 21.72%, 21.72%, and 32.3%, respectively. Moreover, Table 2 further implies the effectiveness of Semi-PNMF compared with CNMF, SCNMF, TSVM, and LDS in terms of both sensitivity and specificity. For completeness, we list their definitions as follows:
(16)
and
(17)
where TP, TN, FP, and FN denote the number of true positive, true negative, false positive and false negative samples, respectively.
The number of labeled examples is an important factor affecting the performance of semi-supervised learning methods. Hence, it is very necessary to observe the classification accuracy of Semi-PNMF under different numbers (1–6) of labeled samples in each class. Here, we randomly select different numbers of examples from each class as labeled examples and regard the rest as unlabeled. For fair comparison, we independently conduct 100 individual experiment trails to remove the effect of randomness.
Fig 8 compares the average accuracy of CNMF, SCNMF, TSVM, LDS, and Semi-PNMF under different numbers of labeled samples for each class. It also shows that Semi-PNMF achieves the highest accuracy and takes on an increasing tendency with the rise in the number of labeled samples.
Acute Leukemia Dataset
We also conduct a cancer classification experiment to verify the classification performance of Semi-PNMF compared with low density separation (LDS, [14]), transductive SVM (TSVM, [16]), constrained NMF (CNMF, [24]), and soft-constrained NMF (SCNMF, [25]) on another popular dataset, i.e., the Acute Leukemia dataset [36]. This dataset comes from Gene Expression Omnibus (http://www.ncbi.nlm.nih.gov/geo/query/acc.cgi?acc=GSE13159), and can also be downloaded from the website: https://zenodo.org/record/21712. We replace the unavailable entries of this dataset with the average values of their k-nearest neighbor elements. This dataset consists of 2,096 samples along with 54,675 probes in total. This dataset contains different cancer subtypes of the acute leukemia and thus is not suited for cancer classification in contrast with the GCM dataset. Table 3 gives a brief description of this dataset. Then, we feed this dataset to all the compared methods.
For Semi-PNMF, we set two parameters α = 0.2, and β = 0.01. For the traditional semi-supervised learning methods, we adopt the same configurations as the above subsection. The cross-validation process of the above subsection is repeatedly conducted to evaluate the compared methods on this dataset. As shown in Figs 9 to 13, the confusion matrix of the predicted results of Semi-PNMF, CNMF, SCNMF, LDS and TSVM are reported in detail. Each column denotes how many unlabeled samples are assigned to each cancer subtype, while each row signifies the number of unlabeled samples affiliated to the real tumor subtype. Each color not only represents a specific cancer subtype but also highlights the correct prediction results, i.e., the diagonal elements of the confusion matrix.
Matrix delineates distribution of actual compared with predicted class membership for multiclass cancer prediction on the Acute Leukemia dataset.
Matrix delineates distribution of actual compared with predicted class membership for multiclass cancer prediction on the Acute Leukemia dataset.
Matrix delineates distribution of actual compared with predicted class membership for multiclass cancer prediction on the Acute Leukemia dataset.
Matrix delineates distribution of actual compared with predicted class membership for multiclass cancer prediction on the Acute Leukemia dataset.
Matrix delineates distribution of actual compared with predicted class membership for multiclass cancer prediction on the Acute Leukemia dataset.
Figs 9 to 13 imply that Semi-PNMF can identify different tumor types more accurately than the representative methods. Semi-PNMF achieves the highest total classification accuracy compared with CNMF, SCNMF, TSVM and LDS in terms of the prediction results in the confusion matrix. Moreover, Table 4 also indicates that Semi-PNMF consistently outperforms the compared methods on eighteen cancer subtypes in terms of both sensitivity and specificity. In summary, these results suggest the effectiveness of Semi-PNMF in cancer classification.
The number of the labeled samples is an important factor affecting the performance of semi-supervised learning methods. Hence, it is very necessary to observe the classification accuracy of Semi-PNMF under different numbers (1–6) of labeled samples in each class. Here, we randomly select different numbers of examples from each class as labeled examples and regard the rest as unlabeled. Then, we independently conduct 10 individual experiment trails to remove the effect of randomness.
Fig 14 compares the average accuracy of CNMF, SCNMF, TSVM, LDS, and Semi-PNMF under different numbers of labeled samples for each class. It also shows that Semi-PNMF achieves the highest accuracy and has an increasing tendency with the rise in the number of labeled samples.
Discussion
This paper proposes the semi-supervised PNMF method (Semi-PNMF), which incorporates two types of constraints as well as the auxiliary basis to boost PNMF. Particularly, Semi-PNMF utilizes the linear combination of examples to approximate the cluster centroids such that the cluster centroids have more powerful representative ability. To effectively indicate the classes of unlabeled samples, Semi-PNMF enforces the coefficients of labeled samples to approach their labels, meanwhile representing the unlabeled samples using the identical cluster centroid. To optimize Semi-PNMF, we devised the multiplicative update rule (MUR) to establish the convergence guarantee. Experiments of cancer classification on two real-world datasets show that Semi-PNMF outperforms the representative methods in terms of quantity.
Recently, Bayesian methods that incorporate both sparsity and a large number of covariates in the model have been extensively used for parameter estimation and classification in data sets compared to small sample sizes such as gene expression data [39–41]. They also improve model accuracy by introducing a slight bias in the model [40]. In future works, we can borrow from the merits of Bayesian methods to further improve the classification performance of Semi-PNMF for a large-scale dataset. Semi-PNMF has provided a flexible framework for learning methods in cancer data processing and can be utilized in other applications such as cancer recurrence [42, 43].
Author Contributions
Conceived and designed the experiments: XZ NG. Performed the experiments: XZ ZJ. Analyzed the data: XZ NG. Contributed reagents/materials/analysis tools: XZ ZJ NG XQ ZL. Wrote the paper: XZ NG.
References
- 1. Golub TR, Slonim DK, Tamayo P, Huard C, Gaasenbeek M, Mesirov JP, et al. Molecular classification of cancer: class discovery and class prediction by gene expression monitoring. science. 1999;286(5439):531–537. pmid:10521349
- 2. Alon U, Barkai N, Notterman DA, Gish K, Ybarra S, Mack D, et al. Broad patterns of gene expression revealed by clustering analysis of tumor and normal colon tissues probed by oligonucleotide arrays. Proceedings of the National Academy of Sciences. 1999;96(12):6745–6750.
- 3. Nguyen DV, Rocke DM. Multi-class cancer classification via partial least squares with gene expression profiles. Bioinformatics. 2002;18(9):1216–1226. pmid:12217913
- 4. Alizadeh AA, Eisen MB, Davis RE, Ma C, Lossos IS, Rosenwald A, et al. Distinct types of diffuse large B-cell lymphoma identified by gene expression profiling. Nature. 2000;403(6769):503–511. pmid:10676951
- 5. Perou CM, Sørlie T, Eisen MB, van de Rijn M, Jeffrey SS, Rees CA, et al. Molecular portraits of human breast tumours. Nature. 2000;406(6797):747–752. pmid:10963602
- 6. Network CGA, et al. Comprehensive molecular portraits of human breast tumours. Nature. 2012;490(7418):61–70.
- 7. Tamayo P, Slonim D, Mesirov J, Zhu Q, Kitareewan S, Dmitrovsky E, et al. Interpreting patterns of gene expression with self-organizing maps: methods and application to hematopoietic differentiation. Proceedings of the National Academy of Sciences. 1999;96(6):2907–2912.
- 8. Ramaswamy S, Tamayo P, Rifkin R, Mukherjee S, Yeang CH, Angelo M, et al. Multiclass cancer diagnosis using tumor gene expression signatures. Proceedings of the National Academy of Sciences. 2001;98(26):15149–15154.
- 9. Bicciato S, Luchini A, Di Bello C. PCA disjoint models for multiclass cancer analysis using gene expression data. Bioinformatics. 2003;19(5):571–578. pmid:12651714
- 10. Tan Y, Shi L, Tong W, Hwang GG, Wang C. Multi-class tumor classification by discriminant partial least squares using microarray gene expression data and assessment of classification models. Computational Biology and Chemistry. 2004;28(3):235–243. pmid:15261154
- 11. Tan Y, Shi L, Tong W, Wang C. Multi-class cancer classification by total principal component regression (TPCR) using microarray gene expression data. Nucleic acids research. 2005;33(1):56–65. pmid:15640445
- 12.
Rui, X, Anagnostopoulos, G. Multiclass Cancer Classification Using Semisupervised Ellipsoid ARTMAP and Particle Swarm Optimization with Gene Expression Data. 2007;.
- 13. Shi M, Zhang B. Semi-supervised learning improves gene expression-based prediction of cancer recurrence. Bioinformatics. 2011;27(21):3017–3023. pmid:21893520
- 14.
Chapelle O, Zien A. Semi-supervised classification by low density separation. In: Proceedings of the 10-th International Workshop on Artificial Intelligence and Statistics, 2005. p. 57–64.
- 15. Maulik U, Mukhopadhyay A, Chakraborty D. Gene-expression-based cancer subtypes prediction through feature selection and transductive SVM. IEEE Transactions on Biomedical Engineering. 2013;60(4):1111–1117. pmid:23095982
- 16.
Vapnik VN, Vapnik V. Statistical learning theory. Wiley New York; 1998.
- 17. Lee DD, Seung HS. Learning the parts of objects by non-negative matrix factorization. Nature. 1999;401(6755):788–791. pmid:10548103
- 18.
Lee DD, Seung HS. Algorithms for non-negative matrix factorization. In: Advances in neural information processing systems; 2001. p. 556–562.
- 19. Guan N, Tao D, Luo Z, Yuan B. NeNMF: an optimal gradient method for nonnegative matrix factorization. IEEE Transactions on Signal Processing. 2012;60(6):2882–2898.
- 20. Cho Y, Saul LK. Nonnegative Matrix Factorization for Semi-supervised Dimensionality Reduction. arXiv preprint arXiv:11123714. 2011.
- 21. Yang Z, Oja E. Linear and nonlinear projective nonnegative matrix factorization. IEEE Transactions on Neural Networks. 2010;21(5):734–749. pmid:20350841
- 22.
Guan N, Huang X, Lan L, Luo Z, Zhang X. Graph based semi-supervised non-negative matrix factorization for document clustering. In: 2012 IEEE 11th International Conference on Machine Learning and Applications (ICMLA); 2012. p.404–408.
- 23. Shahnaz F, Berry MW, Pauca VP, Plemmons RJ. Document clustering using nonnegative matrix factorization. Information Processing & Management. 2006;42(2):373–386.
- 24. Liu H, Wu Z, Li X, Cai D, Huang TS. Constrained nonnegative matrix factorization for image representation. IEEE Transactions on Pattern Analysis and Machine Intelligence. 2012;34(7):1299–1311.
- 25.
Lan L, Guan N, Zhang X, Tao D, Luo Z. Soft-constrained nonnegative matrix factorization via normalization. In: 2014 IEEE International Joint Conference on Neural Networks (IJCNN); 2014. p. 3025–3030.
- 26.
Zhang X, Guan N, Lan L, Tao D, Luo Z. Box-constrained projective nonnegative matrix factorization via augmented Lagrangian method. In: 2014 IEEE International Joint Conference on Neural Networks (IJCNN); 2014. p. 1900–1906.
- 27. Lee CM, Mudaliar MA, Haggart D, Wolf CR, Miele G, Vass JK, et al. Simultaneous non-negative matrix factorization for multiple large scale gene expression datasets in toxicology. PloS ONE. 2012;7(12):e48238. pmid:23272042
- 28. Taslaman L, Nilsson B. A framework for regularized non-negative matrix factorization, with application to the analysis of gene expression data. PloS ONE. 2012;7(11):e46331. pmid:23133590
- 29. Zhang S, Li Q, Liu J, Zhou XJ. A novel computational framework for simultaneous integration of multiple types of genomic data to identify microRNA-gene regulatory modules. Bioinformatics. 2011;27(13):i401–i409. pmid:21685098
- 30. Fogel P, Young SS, Hawkins DM, Ledirac N. Inferential, robust non-negative matrix factorization analysis of microarray data. Bioinformatics. 2007;23(1):44–49. pmid:17092989
- 31. Brunet JP, Tamayo P, Golub TR, Mesirov JP. Metagenes and molecular pattern discovery using matrix factorization. Proceedings of the national academy of sciences. 2004;101(12):4164–4169.
- 32.
Li SZ, Hou X, Zhang H, Cheng Q. Learning spatially localized, parts-based representation. In: Proceedings of the 2001 IEEE Computer Society Conference on Computer Vision and Pattern Recognition; 2001. p. 207–212.
- 33. Hoyer PO. Non-negative matrix factorization with sparseness constraints. The Journal of Machine Learning Research. 2004;5:1457–1469.
- 34. Gao Y, Church G. Improving molecular cancer class discovery through sparse non-negative matrix factorization. Bioinformatics. 2005;21(21):3970–3975. pmid:16244221
- 35.
Yuan Z, Oja E. Projective nonnegative matrix factorization for image compression and feature extraction. In: Image Analysis. Springer; 2005. p. 333–342.
- 36. Staunton JE, Slonim DK, Coller HA, Tamayo P, Angelo MJ, Park J, et al. Chemosensitivity prediction by transcriptional profiling. Proceedings of the National Academy of Sciences. 2001;98(19):10787–10792.
- 37.
Zhang H, Yang Z, Oja E. Adaptive multiplicative updates for projective nonnegative matrix factorization. In: Neural Information Processing. Springer; 2012. p. 277–284.
- 38. Yu K, Zhang T, Gong Y. Nonlinear Learning using Local Coordinate Coding. In: Advances in Neural Information Processing Systems. vol. 22; 2009. p. 2223–2231.
- 39. Kyounghwa B, Mallick B. Gene Selection Using a Two-level Hierarchical Bayesian Model. Bioinformatics. 2004;20(18):3423–3430.
- 40. Behrouz M, Deng L, Homayouni R. Application of Sparse Bayesian Generalized Linear Model to Gene Expression Data for Classification of Prostate Cancer Subtypes. Bioinformatics. 2014;20(18):3423–3430.
- 41. Li J, Das K, Fu G, Li R, Wu R. The Bayesian Lasso for Genome-wide Association Studies. Bioinformatics. 2011;27(4):516–523. pmid:21156729
- 42. Shi M, Zhang B. Semi-supervised Learning Improves Gene Expression-based Prediction of Cancer Recurrence. Bioinformatics. 2011;27(21):3017–3023. pmid:21893520
- 43. Hofree M, Shen J, Carter H, Gross A, Ideker T. Network-based Stratification of Tumor Mutations. Nature methods. 2013;10(11):1108–1115. pmid:24037242