Abstract:
|
In this work, we explore the connection between model explainability and model transferability using electrical signal spectrogram data. Transferability of machine learning (ML) models in real-world scientific applications are dependent on understanding how physical features map to model features important in classification and whether the features usage changes with new datasets. While frequently characterized with train/test set accuracies, cross-validation, and more, these approaches do not necessarily provide understandable data-driven reasons as to where and why a model is (or isn’t) transferable. In this work, we take a two-pronged approach that keeps explainability at the forefront. By first training a novel convolutional neural network architecture under data-driven constraints and subsequently utilizing a post hoc explainability technique, Locally Interpretably Model-agnostic Explainability, we can track model feature usage and compare this usage across data sets by building metrics to compare domain shift. Initial experimental results using this technique will be presented.
|