Online Program Home
  My Program

Abstract Details

Activity Number: 414 - Model Building and Selection
Type: Contributed
Date/Time: Tuesday, August 1, 2017 : 2:00 PM to 3:50 PM
Sponsor: Section on Statistical Computing
Abstract #324273 View Presentation
Title: Number of Samples Needed for Model Selection with Confidence
Author(s): Sean Carver*
Companies: American University
Keywords: Model Selection ; Likelihood Ratio Test ; Akaike Information Criterion ; Kullback-Leibler Divergence ; Baseball
Abstract:

A common measure used to quantify the similarity of two models is the Kullback-Leibler divergence, computed from a true model to an alternative model. We propose a different measure: the number of samples needed to correctly reject the alternative model with a given confidence level (e.g. 95%). Our method works as follows: (1) we simulate samples from the true model, (2) for each sample, we compute a log-likelihood ratio (3), we bootstrap and sum the log-likelihood ratios---when this sum is positive, we select the true model, (4) using simple linear regression, we determine the number of terms (i.e. number of samples) needed to make the desired quantile (e.g. 5%) fall at zero. We have tested this method on t-distributions of different degrees of freedom and have confirmed that it gives reasonably consistent results. However, we plan to apply this method to Markov chains, e.g. used for sports statistics like tennis, volleyball, and baseball. For these applications, it may be desirable to have a measure that is easier to interpret than the Kullback-Leibler divergence. How many innings are needed to falsify the model of the Yankees when simulating a model of the Orioles?


Authors who are presenting talks have a * after their name.

Back to the full JSM 2017 program

 
 
Copyright © American Statistical Association