Maxi-Min discriminant analysis via online learning.
Linear Discriminant Analysis (LDA) is an important dimensionality reduction algorithm, but its performance is usually limited on multi-class data. Such limitation is incurred by the fact that LDA actually maximizes the average divergence among classes, whereby similar classes with smaller divergence tend to be merged in the subspace. To address this problem, we propose a novel dimensionality reduction method called Maxi-Min Discriminant Analysis (MMDA). In contrast to the traditional LDA, MMDA attempts to find a low-dimensional subspace by maximizing the minimal (worst-case) divergence among classes. This "minimal" setting overcomes the problem of LDA that tends to merge similar classes with smaller divergence when used for multi-class data. We formulate MMDA as a convex problem and further as a large-margin learning problem. One key contribution is that we design an efficient online learning algorithm to solve the involved problem, making the proposed method applicable to large scale data. Experimental results on various datasets demonstrate the efficiency and the efficacy of our proposed method against five other competitive approaches, and the scalability to the data with thousands of classes.
Duke Scholars
Published In
DOI
EISSN
ISSN
Publication Date
Volume
Start / End Page
Related Subject Headings
- Pattern Recognition, Automated
- Discrimination Learning
- Discriminant Analysis
- Artificial Intelligence & Image Processing
- Artificial Intelligence
- Algorithms
- 4905 Statistics
- 4611 Machine learning
- 4602 Artificial intelligence
Citation
Published In
DOI
EISSN
ISSN
Publication Date
Volume
Start / End Page
Related Subject Headings
- Pattern Recognition, Automated
- Discrimination Learning
- Discriminant Analysis
- Artificial Intelligence & Image Processing
- Artificial Intelligence
- Algorithms
- 4905 Statistics
- 4611 Machine learning
- 4602 Artificial intelligence