Content area
Abstract
In recent months, researchers, government bodies, and the media have claimed that a ‘right to explanation’ of decisions made by automated and artificially intelligent algorithmic systems is legally mandated by the forthcoming European Union General Data Protection Regulation 2016/679 (GDPR). The right to explanation is viewed as a promising mechanism in the broader pursuit by government and industry for accountability and transparency in algorithms, artificial intelligence, robotics, and other automated systems. However, there are several reasons to doubt the existence, scope, and feasibility of a ‘right to explanation’ of automated decisions. In this article, we examine the legal status of the ‘right to explanation’ in the GDPR, and identify several barriers undermining its implementation. We argue that the GDPR does not, in its current form, implement a right to explanation, but rather what we term a limited ‘right to be informed’.





