10.1145/1390156.1390277acmotherconferencesArticle/Chapter ViewAbstractPublication PagesicpsprocConference Proceedings
research-article

Expectation-maximization for sparse and non-negative PCA

ABSTRACT

We study the problem of finding the dominant eigenvector of the sample covariance matrix, under additional constraints on the vector: a cardinality constraint limits the number of non-zero elements, and non-negativity forces the elements to have equal sign. This problem is known as sparse and non-negative principal component analysis (PCA), and has many applications including dimensionality reduction and feature selection. Based on expectation-maximization for probabilistic PCA, we present an algorithm for any combination of these constraints. Its complexity is at most quadratic in the number of dimensions of the data. We demonstrate significant improvements in performance and computational efficiency compared to other constrained PCA algorithms, on large data sets from biology and computer vision. Finally, we show the usefulness of non-negative sparse PCA for unsupervised feature selection in a gene clustering task.

References

  1. Armstrong, S., Staunton, J., Silverman, L., Pieters, R., den Boer, M., Minden, M., Sallan, S., Lander, E., Golub, T., & Korsmeyer, S. (2002). MLL translocations specify a distinct gene expression profile that distinguishes a unique leukemia. Nature Genetics, 30, 41--47.Google ScholarGoogle ScholarCross RefCross Ref
  2. Cadima, J., & Jolliffe, I. (1995). Loadings and correlations in the interpretation of principal components. Applied Statistics, 203--214.Google ScholarGoogle Scholar
  3. d'Aspremont, A., Bach, F., & El Ghaoui, L. (2007). Full regularization path for sparse principal component analysis. Proceedings of the International Conference on Machine Learning. Google ScholarGoogle ScholarDigital LibraryDigital Library
  4. Horst, R., Pardalos, P., & Thoai, N. (2000). Introduction to global optimization. Kluwer Acad. Publ. Google ScholarGoogle ScholarDigital LibraryDigital Library
  5. Moghaddam, B., Weiss, Y., & Avidan, S. (2006). Spectral bounds for sparse PCA: Exact and greedy algorithms. Advances in Neural Information Processing Systems.Google ScholarGoogle Scholar
  6. Roweis, S. (1998). EM algorithms for PCA and sensible PCA. Advances in Neural Information Processing Systems.Google ScholarGoogle Scholar
  7. Sha, F., Lin, Y., Saul, L., & Lee, D. (2007). Multiplicative Updates for Nonnegative Quadratic Programming. Neural Computation, 19, 2004--2031. Google ScholarGoogle ScholarDigital LibraryDigital Library
  8. Sriperumbudur, B., Torres, D., & Lanckriet, G. (2007). Sparse eigen methods by d.c. programming. Proceedings of the International Conference on Machine Learning. Google ScholarGoogle ScholarDigital LibraryDigital Library
  9. Sung, K.-K. (1996). Learning and example selection for object and pattern recognition. Doctoral dissertation, MIT, Artificial Intelligence Laboratory and Center for Biological and Computational Learning, Cambridge, MA. Google ScholarGoogle ScholarDigital LibraryDigital Library
  10. Tibshirani, R. (1996). Regression shrinkage and selection via the LASSO. Journal of the Royal statistical society, series B, 58, 267--288.Google ScholarGoogle ScholarCross RefCross Ref
  11. Tipping, M., & Bishop, C. (1999). Probabilistic principal component analysis. Journal of the Royal Statistical Society, Series B, 21, 611--622. Google ScholarGoogle ScholarDigital LibraryDigital Library
  12. Varshavsky, R., Gottlieb, A., Linial, M., & Horn, D. (2006). Novel Unsupervised Feature Filtering of Biological Data. Bioinformatics, 22. Google ScholarGoogle ScholarDigital LibraryDigital Library
  13. Zass, R., & Shashua, A. (2006). Nonnegative sparse PCA. Advances in Neural Information Processing Systems.Google ScholarGoogle Scholar
  14. Zou, H., Hastie, T., & Tibshirani, R. (2004). Sparse principal component analysis. Journal of Computational and Graphical Statistics.Google ScholarGoogle Scholar

Index Terms

  1. Expectation-maximization for sparse and non-negative PCA

              Comments

              Login options

              Check if you have access through your login credentials or your institution to get full access on this article.

              Sign in
              • Published in

                ACM Other conferences cover image
                ICML '08: Proceedings of the 25th international conference on Machine learning
                July 2008
                1310 pages
                ISBN:9781605582054
                DOI:10.1145/1390156

                Copyright © 2008 ACM

                Publisher

                Association for Computing Machinery

                New York, NY, United States

                Publication History

                • Published: 5 July 2008

                Permissions

                Request permissions about this article.

                Request Permissions

                Qualifiers

                • research-article

                Acceptance Rates

                ICML '08 Paper Acceptance Rate 158 of 583 submissions, 27%
                Overall Acceptance Rate 448 of 1,653 submissions, 27%

              PDF Format

              View or Download as a PDF file.

              PDF

              eReader

              View online with eReader.

              eReader
              About Cookies On This Site

              We use cookies to ensure that we give you the best experience on our website.

              Learn more

              Got it!