Nonnegative Sparse PCA

Ron Zass*, Amnon Shashua

*Corresponding author for this work

Research output: Chapter in Book/Report/Conference proceedingConference contributionpeer-review

133 Scopus citations


We describe a nonnegative variant of the "Sparse PCA" problem. The goal is to create a low dimensional representation from a collection of points which on the one hand maximizes the variance of the projected points and on the other uses only parts of the original coordinates, and thereby creating a sparse representation. What distinguishes our problem from other Sparse PCA formulations is that the projection involves only nonnegative weights of the original coordinates - a desired quality in various fields, including economics, bioinformatics and computer vision. Adding nonnegativity contributes to sparseness, where it enforces a partitioning of the original coordinates among the new axes. We describe a simple yet efficient iterative coordinate-descent type of scheme which converges to a local optimum of our optimization criteria, giving good results on large real world datasets.

Original languageAmerican English
Title of host publicationAdvances in Neural Information Processing Systems 19 - Proceedings of the 2006 Conference
Number of pages8
StatePublished - 2007
Event20th Annual Conference on Neural Information Processing Systems, NIPS 2006 - Vancouver, BC, Canada
Duration: 4 Dec 20067 Dec 2006

Publication series

NameAdvances in Neural Information Processing Systems
ISSN (Print)1049-5258


Conference20th Annual Conference on Neural Information Processing Systems, NIPS 2006
CityVancouver, BC


Dive into the research topics of 'Nonnegative Sparse PCA'. Together they form a unique fingerprint.

Cite this