10.1145/1291233.1291297acmconferencesArticle/Chapter ViewAbstractPublication PagesmmConference Proceedingsconference-collections
Article

Feature selection using principal feature analysis

Authors Info & Claims
Published:29 September 2007Publication History

ABSTRACT

Dimensionality reduction of a feature set is a common preprocessing step used for pattern recognition and classification applications. Principal Component Analysis (PCA) is one of the popular methods used, and can be shown to be optimal using different optimality criteria. However, it has the disadvantage that measurements from all the original features are used in the projection to the lower dimensional space. This paper proposes a novel method for dimensionality reduction of a feature set by choosing a subset of the original features that contains most of the essential information, using the same criteria as PCA. We call this method Principal Feature Analysis (PFA). The proposed method is successfully applied for choosing the principal features in face tracking and content-based image retrieval (CBIR) problems. Automated annotation of digital pictures has been a highly challenging problem for computer scientists since the invention of computers. The capability of annotating pictures by computers can lead to breakthroughs in a wide range of applications including Web image search, online picture-sharing communities, and scientific experiments. In our work, by advancing statistical modeling and optimization techniques, we can train computers about hundreds of semantic concepts using example pictures from each concept. The ALIPR (Automatic Linguistic Indexing of Pictures - Real Time) system of fully automatic and high speed annotation for online pictures has been constructed. Thousands of pictures from an Internet photo-sharing site, unrelated to the source of those pictures used in the training process, have been tested. The experimental results show that a single computer processor can suggest annotation terms in real-time and with good accuracy.

References

  1. Lisboa, P. J. G., Mehri-Dehnavi, R. Sensitivity Methods for Variable Selection Using the MLP. International Workshop on Neural Networks for Identification, Control, Robotics and Signal/Image, 1996, 330--338. Google ScholarGoogle ScholarDigital LibraryDigital Library
  2. Lin, T.-S., Meador, J. Statistical Feature Extraction and Selection for IC Test Pattern Analysis. Circuits and systems, vol 1., 1992, 391--394.Google ScholarGoogle Scholar
  3. Hocking, R. R. Development in Linear Regression Methodology: 1959-1982. Technometrics, vol. 25, 1983, 219--249.Google ScholarGoogle Scholar
  4. McCabe, G. P. Principal Variables. Technometrics, vol. 26, 1984, 127--134.Google ScholarGoogle Scholar
  5. Jolliffe, I. T. Principal Component Analysis. Springer-Verlag, New-York, 1986.Google ScholarGoogle Scholar
  6. Krzanowski, W. J. Selection of Variables to Preserve Multivariate Data Structure, Using Principal Component Analysis. Applied Statistics - Journal of the Royal Statistical Society Series C, vol. 36, 1987, 22--33.Google ScholarGoogle Scholar
  7. Krzanowski, W. J. A Stopping Rule for structure - Preserving Variable Selection. Statistics and Computing, March vol. 6, 1996, 51--56.Google ScholarGoogle Scholar
  8. Gower, J. C., Statistical Methods of Comparing Different Multivariate Analyses of the Same Data, Mathematics in the Archaeological and Historical Sciences, University Press, Edinburgh, 1971, 138--149.Google ScholarGoogle Scholar

Index Terms

  1. Feature selection using principal feature analysis

      Recommendations

      Comments

      Login options

      Check if you have access through your login credentials or your institution to get full access on this article.

      Sign in

      PDF Format

      View or Download as a PDF file.

      PDF

      eReader

      View online with eReader.

      eReader
      About Cookies On This Site

      We use cookies to ensure that we give you the best experience on our website.

      Learn more

      Got it!