A responsible machine learning workflow with focus on interpretable models, post-hoc explanation, and discrimination testing

  • Navdeep Gill
  • , Patrick Hall
  • , Kim Montgomery
  • , Nicholas Schmidt

Research output: Contribution to journalArticlepeer-review

Abstract

This manuscript outlines a viable approach for training and evaluating machine learning systems for high-stakes, human-centered, or regulated applications using common Python programming tools. The accuracy and intrinsic interpretability of two types of constrained models, monotonic gradient boosting machines and explainable neural networks, a deep learning architecture well-suited for structured data, are assessed on simulated data and publicly available mortgage data. For maximum transparency and the potential generation of personalized adverse action notices, the constrained models are analyzed using post-hoc explanation techniques including plots of partial dependence and individual conditional expectation and with global and local Shapley feature importance. The constrained model predictions are also tested for disparate impact and other types of discrimination using measures with long-standing legal precedents, adverse impact ratio, marginal effect, and standardized mean difference, along with straightforward group fairness measures.

Original languageEnglish
Article number137
JournalInformation (Switzerland)
Volume11
Issue number3
DOIs
StatePublished - Mar 1 2020
Externally publishedYes

Bibliographical note

Publisher Copyright:
© 2020 by the authors. Licensee MDPI, Basel, Switzerland.

ASJC Scopus Subject Areas

  • Information Systems

Keywords

  • Deep learning
  • Disparate impact
  • Explanation
  • Fairness
  • Gradient boosting machine
  • Interpretable
  • Machine learning
  • Neural network
  • Python

Fingerprint

Dive into the research topics of 'A responsible machine learning workflow with focus on interpretable models, post-hoc explanation, and discrimination testing'. Together they form a unique fingerprint.

Cite this