Manipulating and measuring model interpretability
Forough Poursabzi-Sangdeh argues that to understand interpretability, we need to bring humans in the loop and run human-subject experiments. She describes a set of controlled user experiments in which researchers manipulated various design factors in models that are commonly thought to make them more or less interpretable and measured their influence on users behavior.
Talk Title | Manipulating and measuring model interpretability |
Speakers | Forough Poursabzi-Sangdeh (Microsoft Research NYC) |
Conference | O’Reilly Artificial Intelligence Conference |
Conf Tag | Put AI to Work |
Location | New York, New York |
Date | April 16-18, 2019 |
URL | Talk Page |
Slides | Talk Slides |
Video | |
Machine learning is increasingly used to make decisions that affect people’s lives in critical domains like criminal justice, fair lending, and medicine. While most of the research in machine learning focuses on improving the performance of models on held-out datasets, this is seldom enough to convince end users that these models are trustworthy and reliable in the wild. To address this problem, a new line of research has emerged that focuses on developing interpretable machine learning methods and helping end users make informed decisions. Despite the growing body of work in developing interpretable models, there is still no consensus on the definition and quantification of interpretability. Forough Poursabzi-Sangdeh argues that to understand interpretability, we need to bring humans in the loop and run human-subject experiments. Forough approaches the problem of interpretability from an interdisciplinary perspective built on decades of research in psychology, cognitive science, and social science to understand human behavior and trust. She describes a set of controlled user experiments in which researchers manipulated various design factors in models that are commonly thought to make them more or less interpretable and measured their influence on users’ behavior. The findings emphasize the importance of studying how models are presented to people and empirically verifying that interpretable models achieve their intended effects on end users.