This project considers whether replicable validation of a machine learning (ML) model’s efficacy should be emphasized over an intractable requirement of interpretability, and if such interpretability may in certain situations prove deceptive or deleterious. We argue why interpretability should have primacy alongside empiricism for several reasons: first, if machine learning models are beginning to render some of the high-risk healthcare decisions instead of clinicians, these models pose a novel medico-legal and ethical frontier that is incompletely addressed by current methods of appraising medical interventions like pharmacological therapies; second, a number of judicial precedents underpinning medical liability and negligence are compromised when ’autonomous’ ML recommendations are considered to be en par with human instruction in specific contexts; third, explainable algorithms may be more amenable to the ascertainment and minimization of biases, with repercussions for racial equity as well as scientific reproducibility and generalizability. We conclude with some reasons for the ineludible importance of interpretability, such as the establishment of trust, in overcoming perhaps the most difficult challenge ML will face in a high-stakes environment like healthcare: professional and public acceptance.
artificial intelligence, machine learning, healthcare, big data, medical ethics
Machine learning offers truly unprecedented diagnostic and prognostic opportunities as medicine endeavors to become more personalized and precise than ever before. Looking towards the clinical deployment of machine learning in healthcare, we explore whether ethical and judicial precedents would be better served if the growing body of ‘empirical validation’ studies were to prioritize, publish and share interpretable machine learning models.