Online Program Home
  My Program

All Times EDT

Abstract Details

Activity Number: 177 - Statistical Modeling of Lifetime Data: LiDS Section Student Award Session
Type: Contributed
Date/Time: Tuesday, August 4, 2020 : 10:00 AM to 2:00 PM
Sponsor: Lifetime Data Science Section
Abstract #313136
Title: BigSurvSGD: Big Survival Data Analysis via Stochastic Gradient Descent
Author(s): Aliasghar Tarkhan* and Noah Simon
Companies: Department of Biostatistics, University of Washington and University of Washington
Keywords: Survival Analysis; Cox Proportional Hazards model; Stochastic gradient descent; Computational stability; Big Data; Deep learning
Abstract:

To assess the connection between features of a patient and a time-to-event outcome (e.g., disease progression), it is common to assume a proportional hazards model, and fit a proportional hazards regression (or Cox regression). To fit this model for moderate-sized datasets, a log-concave objective function known as the partial likelihood is maximized using an efficient Newton-Raphson algorithm. However, in large datasets this approach has two issues: 1) The computational tricks that leverage structure can lead to computational instability; 2) The objective does not naturally decouple: Thus, if the dataset does not fit in memory, the model can be very computationally expensive to fit. This additionally means that the objective is not directly amenable to stochastic gradient-based optimization methods. To overcome these issues, we propose a simple, new framing of proportional hazards regression: This results in an objective function that is amenable to stochastic gradient descent. We show that this simple modification allows us to efficiently fit survival models with very large datasets and it also facilitates training complex, e.g., neural-network-based, models with survival data.


Authors who are presenting talks have a * after their name.

Back to the full JSM 2020 program