Divergence Maximizing Linear Projection for Supervised Dimension Reduction

Biao Chen, Joshua Kortje

Research output: Contribution to journalArticlepeer-review

Abstract

This paper proposes two linear projection methods for supervised dimension reduction using only first- and second-order statistics. The methods, each catering to a different parameter regime, are derived under the general Gaussian model by maximizing the Kullback-Leibler divergence between the two classes in the projected sample for a binary classification problem. They subsume existing linear projection approaches developed under simplifying assumptions of Gaussian distributions, such as these distributions might share an equal mean or covariance matrix. As a by-product, we establish that the multi-class linear discriminant analysis, a celebrated method for classification and supervised dimension reduction, is provably optimal for maximizing pairwise Kullback-Leibler divergence when the Gaussian populations share an identical covariance matrix. For the case when the Gaussian distributions share an equal mean, we establish conditions under which the optimal subspace remains invariant regardless of how the Kullback-Leibler divergence is defined, despite the asymmetry of the divergence measure itself. Such conditions encompass the classical case of signal plus noise, where both signal and noise have zero mean and arbitrary covariance matrices. Experiments are conducted to validate the proposed solutions, demonstrate their superior performance over existing alternatives, and illustrate the procedure for selecting the appropriate linear projection solution.

Original languageEnglish (US)
Pages (from-to)2104-2115
Number of pages12
JournalIEEE Transactions on Information Theory
Volume71
Issue number3
DOIs
StatePublished - 2025

Keywords

  • Kullback-Leibler divergence
  • classification
  • linear discriminate analysis
  • linear projection
  • supervised dimension reduction

ASJC Scopus subject areas

  • Information Systems
  • Computer Science Applications
  • Library and Information Sciences

Fingerprint

Dive into the research topics of 'Divergence Maximizing Linear Projection for Supervised Dimension Reduction'. Together they form a unique fingerprint.

Cite this