Interpretable machine learning

One major challenge for data science is to ensure that models do more good than harm. Along the way, we must ensure that models interpretations are valid, understand when they yield valid causal insights, and identify potential harms for subgroups. Our lab focuses on building tools for interpretable machine learning, which we view as a key component of trustworthy data science. These include powerful predictive models that are also interpretable, and improved methods to handle missing-not-at-random data and informative missing values. These methods have important implications in healthcare, where they have already produced novel and actionable clinical insights in cardiology.

Talks

Software

  • ControlBurn: interpretable feature selection using (tree-based) generalized additive models

Papers

The Missing Indicator Method: From Low to High Dimensions
V. N. Mike, T. Bosschieter, R. Halpin-Gregorio, and M. Udell
29th SIGKDD Conference on Knowledge Discovery and Data Mining - Applied Data Science Track, 2023
[arxiv][bib]

Interpretable Survival Analysis for Heart Failure Risk Prediction
M. Van Ness, T. Bosschieter, N. Din, A. Ambrosy, A. Singh, and M. Udell
Machine Learning for Health (ML4H), 2023
[arxiv][url][bib]

Data-Efficient and Interpretable Tabular Anomaly Detection
C. Chang, J. Yoon, S. Arik, M. Udell, and T. Pfister
arXiv preprint arXiv:2203.02034, 2023
[url][bib]

ControlBurn: Nonlinear Feature Selection with Sparse Tree Ensembles
B. Liu, M. Xie, H. Yang, and M. Udell
2022
[url][bib]

ControlBurn: Feature Selection by Sparse Forests
B. Liu, M. Xie, and M. Udell
ACM SIGKDD Conference on Knowledge Discovery and Data Mining (KDD), 2021
[arxiv][url][slides][bib]

Impact of Accuracy on Model Interpretations
B. Liu and M. Udell
2020
[arxiv][url][bib]

``Why should you trust my explanation?'' Understanding uncertainty in LIME explanations
Y. Zhang, K. Song, Y. Sun, S. Tan, and M. Udell
ICML Workshop AI for Social Good, 2019
[arxiv][bib]

Fairness Under Unawareness: Assessing Disparity When Protected Class Is Unobserved
J. Chen, N. Kallus, X. Mao, G. Svacha, and M. Udell
FAT*: Conference on Fairness, Accountability, and Transparency, 2019
[arxiv][pdf][slides][bib]

Causal Inference with Noisy and Missing Covariates via Matrix Factorization
N. Kallus, X. Mao, and M. Udell
Advances in Neural Information Processing Systems, 2018
[arxiv][code][bib]