Activity Number:
|
181
- SPEED: Statistical Learning and Data Science Speed Session 1, Part 2
|
Type:
|
Contributed
|
Date/Time:
|
Monday, July 29, 2019 : 10:30 AM to 11:15 AM
|
Sponsor:
|
Section on Statistical Learning and Data Science
|
Abstract #307522
|
|
Title:
|
Model Selection for Mixture of Experts Using Group Fused Lasso
|
Author(s):
|
Tuan Do* and Karl Gregory
|
Companies:
|
University of South Carolina and University of South Carolina
|
Keywords:
|
Mixture of experts;
fused lasso;
blockwise coordinate descent;
duality;
convex optimization
|
Abstract:
|
The mixture of experts, or mixture of learners in general, is a popular and powerful machine learning model in which each expert learns to handle a different region of the covariate space. However, it is crucial to choose an appropriate number of experts to avoid overfitting or underfitting. We introduce a group fused lasso term to the model with the goal of making the coefficients of experts and gating networks closer together. By varying the strength of the penalization, we can avoid overspecialization of each expert and choose the optimal number of experts at the same time. An efficient algorithm to optimize the problem is developed using block-wise coordinate descent in the dual counterpart. Numerical results on simulated and real world datasets show that the penalized model outperforms the unpenalized one and performs on par with many well-known machine learning models.
|
Authors who are presenting talks have a * after their name.