Online Program

Return to main conference page
Saturday, October 20
Knowledge
Sat, Oct 20, 10:00 AM - 11:30 AM
Rosewood
Celebrating Our Technical Contributions

Supervised Dimensionality Reduction for Exponential Family Data (304862)

*Yoonkyung Lee, The Ohio State University 
Andrew Landgraf, Battelle Memorial Institute  

Keywords: dimension reduction, exponential family, generalized linear model, latent factors, PCA

Supervised dimensionality reduction techniques, such as partial least squares and supervised principal components, are powerful tools for making predictions with a large number of variables. The implicit squared error terms in the objectives, however, make it less attractive to non-Gaussian data, either in the covariates or the responses. Drawing on a connection between partial least squares and the Gaussian distribution, we show how partial least squares can be extended to other members of the exponential family - similar to the generalized linear model - for both the covariates and the responses. Unlike previous attempts, our extension gives latent variables which are easily interpretable as linear functions of the data and is computationally efficient. In particular, it does not require additional optimization for the scores of new observations and therefore predictions can be made in real time.