Abstract

We aim to produce predictive models that are not only accurate, but are also\ninterpretable to human experts. Our models are decision lists, which consist of\na series of if...then... statements (e.g., if high blood pressure, then stroke)\nthat discretize a high-dimensional, multivariate feature space into a series of\nsimple, readily interpretable decision statements. We introduce a generative\nmodel called Bayesian Rule Lists that yields a posterior distribution over\npossible decision lists. It employs a novel prior structure to encourage\nsparsity. Our experiments show that Bayesian Rule Lists has predictive accuracy\non par with the current top algorithms for prediction in machine learning. Our\nmethod is motivated by recent developments in personalized medicine, and can be\nused to produce highly accurate and interpretable medical scoring systems. We\ndemonstrate this by producing an alternative to the CHADS$_2$ score, actively\nused in clinical practice for estimating the risk of stroke in patients that\nhave atrial fibrillation. Our model is as interpretable as CHADS$_2$, but more\naccurate.\n

Keywords

Machine learningComputer scienceArtificial intelligenceBayesian probabilityFeature (linguistics)Data mining

Affiliated Institutions

Related Publications

Publication Info

Year
2015
Type
article
Volume
9
Issue
3
Citations
743
Access
Closed

Social Impact

Social media, news, blog, policy document mentions

Citation Metrics

743
OpenAlex
55
Influential
469
CrossRef

Cite This

Benjamin Letham, Cynthia Rudin, Tyler H. McCormick et al. (2015). Interpretable classifiers using rules and Bayesian analysis: Building a better stroke prediction model. The Annals of Applied Statistics , 9 (3) . https://doi.org/10.1214/15-aoas848

Identifiers

DOI
10.1214/15-aoas848
arXiv
1511.01644

Data Quality

Data completeness: 84%