ABSTRACT
In machine learning often a tradeoff must be made between accuracy and intelligibility. More accurate models such as boosted trees, random forests, and neural nets usually are not intelligible, but more intelligible models such as logistic regression, naive-Bayes, and single decision trees often have significantly worse accuracy. This tradeoff sometimes limits the accuracy of models that can be applied in mission-critical applications such as healthcare where being able to understand, validate, edit, and trust a learned model is important. We present two case studies where high-performance generalized additive models with pairwise interactions (GA2Ms) are applied to real healthcare problems yielding intelligible models with state-of-the-art accuracy. In the pneumonia risk prediction case study, the intelligible model uncovers surprising patterns in the data that previously had prevented complex learned models from being fielded in this domain, but because it is intelligible and modular allows these patterns to be recognized and removed. In the 30-day hospital readmission case study, we show that the same methods scale to large datasets containing hundreds of thousands of patients and thousands of attributes while remaining intelligible and providing accuracy comparable to the best (unintelligible) machine learning methods.
Supplemental Material
- R. Ambrosino, B. Buchanan, G. Cooper, and M. Fine. The use of misclassification costs to learn rule-based decision support models for cost-effective hospital admission strategies. In Proceedings of the Annual Symp. on Comp. Application in Medical Care, 1995.Google Scholar
- G. Cooper, V. Abraham, C. Aliferis, J. Aronis, B. Buchanan, R. Caruana, M. Fine, J. Janosky, G. Livingston, T. Mitchell, S. Montik, and P. Spirtes. Predicting dire outcomes of patients with community acquired pneumonia. Journal of Biomedical Informatics, 38(5):347--366, 2005. Google ScholarDigital Library
- G. Cooper, C. Aliferis, R. Ambrosino, J. Aronis, B. Buchanan, R. Caruana, M. Fine, C. Glymour, G. Gordon, B. Hanusa, J. Janosky, C. Meek, T. Mitchell, T. Richardson, and P. Spirtes. An evaluation of machine-learning methods for predicting pneumonia mortality. Artificial Intelligence in Medicine, 9(2):107--138, 1997.Google ScholarCross Ref
- T. Hastie and R. Tibshirani. Generalized additive models. Chapman & Hall/CRC, 1990.Google Scholar
- Y. Lou, R. Caruana, and J. Gehrke. Intelligible models for classification and regression. In KDD, 2012. Google ScholarDigital Library
- Y. Lou, R. Caruana, J. Gehrke, and G. Hooker. Accurate intelligible models with pairwise interactions. In KDD, 2013. Google ScholarDigital Library
- S. Wood. Generalized additive models: an introduction with R. CRC Press, 2006. Google ScholarDigital Library
Index Terms
- Intelligible Models for HealthCare: Predicting Pneumonia Risk and Hospital 30-day Readmission
Recommendations
Interpretable Machine Learning in Healthcare
BCB '18: Proceedings of the 2018 ACM International Conference on Bioinformatics, Computational Biology, and Health InformaticsThis tutorial extensively covers the definitions, nuances, challenges, and requirements for the design of interpretable and explainable machine learning models and systems in healthcare. We discuss many uses in which interpretable machine learning ...
A unified approach to interpreting model predictions
NIPS'17: Proceedings of the 31st International Conference on Neural Information Processing SystemsUnderstanding why a model makes a certain prediction can be as crucial as the prediction's accuracy in many applications. However, the highest accuracy for large modern datasets is often achieved by complex models that even experts struggle to interpret, ...
A Survey on Bias and Fairness in Machine Learning
Invited TutorialWith the widespread use of artificial intelligence (AI) systems and applications in our everyday lives, accounting for fairness has gained significant importance in designing and engineering of such systems. AI systems can be used in many sensitive ...
Comments