"The human body is a black box": supporting clinical decision-making with deep learning

Zotero / D&S Group / Top-Level Items 2020-05-31

Type Conference Paper Author Mark Sendak Author Madeleine Clare Elish Author Michael Gao Author Joseph Futoma Author William Ratliff Author Marshall Nichols Author Armando Bedoya Author Suresh Balu Author Cara O'Brien URL https://doi.org/10.1145/3351095.3372827 Series FAT* '20 Place Barcelona, Spain Publisher Association for Computing Machinery Pages 99–109 ISBN 978-1-4503-6936-7 Date January 27, 2020 DOI 10.1145/3351095.3372827 Accessed 2020-05-30 Library Catalog ACM Digital Library Abstract Machine learning technologies are increasingly developed for use in healthcare. While research communities have focused on creating state-of-the-art models, there has been less focus on real world implementation and the associated challenges to fairness, transparency, and accountability that come from actual, situated use. Serious questions remain underexamined regarding how to ethically build models, interpret and explain model output, recognize and account for biases, and minimize disruptions to professional expertise and work cultures. We address this gap in the literature and provide a detailed case study covering the development, implementation, and evaluation of Sepsis Watch, a machine learning-driven tool that assists hospital clinicians in the early diagnosis and treatment of sepsis. Sepsis is a severe infection that can lead to organ failure or death if not treated in time and is the leading cause of inpatient deaths in US hospitals. We, the team that developed and evaluated the tool, discuss our conceptualization of the tool not as a model deployed in the world but instead as a socio-technical system requiring integration into existing social and professional contexts. Rather than focusing solely on model interpretability to ensure fair and accountable machine learning, we point toward four key values and practices that should be considered when developing machine learning to support clinical decision-making: rigorously define the problem in context, build relationships with stakeholders, respect professional discretion, and create ongoing feedback loops with stakeholders. Our work has significant implications for future research regarding mechanisms of institutional accountability and considerations for responsibly designing machine learning systems. Our work underscores the limits of model interpretability as a solution to ensure transparency, accuracy, and accountability in practice. Instead, our work demonstrates other means and goals to achieve FATML values in design and in practice. Proceedings Title Proceedings of the 2020 Conference on Fairness, Accountability, and Transparency Short Title "The human body is a black box"