Understanding from Deep Learning Models in Context

Abstract

This paper places into context how the term model in machine learning (ML) contrasts with traditional usages of scientific models for understanding and we show how direct analysis of an estimator’s learned transformations (specifically, the hidden layers of a deep learning model) can improve understanding of the target phenomenon and reveal how the model organizes relevant information. Specifically, three modes of understanding will be identified, the difference between implementation irrelevance and functionally approximate irrelevance will be disambiguated, and how this distinction impacts potential understanding with these models will be explored. Additionally, by distinguishing between empirical link failures from representational ones, an ambiguity in the concept of link uncertainty will be addressed thus clarifying the role played by scientific background knowledge in enabling understanding with ML

    Similar works