Home Page

Papers

Submissions

News

Editorial Board

Open Source Software

Proceedings (PMLR)

Transactions (TMLR)

Search

Statistics

Login

Frequently Asked Questions

Contact Us



RSS Feed

Regularized Discriminant Analysis, Ridge Regression and Beyond

Zhihua Zhang, Guang Dai, Congfu Xu, Michael I. Jordan; 11(76):2199−2228, 2010.

Abstract

Fisher linear discriminant analysis (FDA) and its kernel extension−kernel discriminant analysis (KDA)−are well known methods that consider dimensionality reduction and classification jointly. While widely deployed in practical problems, there are still unresolved issues surrounding their efficient implementation and their relationship with least mean squares procedures. In this paper we address these issues within the framework of regularized estimation. Our approach leads to a flexible and efficient implementation of FDA as well as KDA. We also uncover a general relationship between regularized discriminant analysis and ridge regression. This relationship yields variations on conventional FDA based on the pseudoinverse and a direct equivalence to an ordinary least squares estimator.

[abs][pdf][bib]       
© JMLR 2010. (edit, beta)