Project

Exploring the ethical consequences of black-box medicine

Code
BOF/STA/201909/033
Duration
01 March 2020 → 31 December 2024
Funding
Regional and community funding: Special Research Fund
Research disciplines
  • Humanities
    • Medical ethics and medical deontology
Keywords
medical ethics Artificial Intelligence health care doctor-patiënt relationship innovation
 
Project description

Autonomy, responsibility and the prospects of explainable AI  

Artificial intelligence (AI) tools in medicine are increasingly opaque. This means that it becomes very hard – or simply impossible – to understand why an AI tool generated a certain outcome given the input it received. Moreover, the most promising techniques such as deep neural networks are the most opaque as well. This makes the ethical concerns related to the resulting black-box all the more pressing. For if a patient cannot understand why the AI tool proposes a certain therapy, how can she decide upon the adequacy of its proposal? In this project we will focus on the ethical consequences of black-box AI tools in medicine in terms of autonomy, informed consent, shared-decision making and responsibility.

Firstly, the literature is undecided with regard to the precise nature of the ethical problems arising from these black-box tools. Hence our first goal is to clarify how specific AI tools might problematize autonomy and related ethical concepts.

Secondly, the nascent scientific field of explainable artificial intelligence (XAI) develops techniques to foster (post-hoc) explanations for what happened in the black-box. Therefore we can ask which specific requirements these explanations should meet in order to solve the ethical problems we identified. In order to do this we will combine the bio-ethical literature on black-box AI, the literature in the philosophy of science on explainable AI and the corresponding scientific literature. Ultimately, we hope that our research will attribute to the development of a fine-grained ethical discussion about black-box AI in health care and constructively guide the development of the scientific field of explainable AI.