Embedding Values in Artificial Intelligence (AI) Systems

Minds and Machines 30 (3):385-409 (2020)
  Copy   BIBTEX

Abstract

Organizations such as the EU High-Level Expert Group on AI and the IEEE have recently formulated ethical principles and (moral) values that should be adhered to in the design and deployment of artificial intelligence (AI). These include respect for autonomy, non-maleficence, fairness, transparency, explainability, and accountability. But how can we ensure and verify that an AI system actually respects these values? To help answer this question, I propose an account for determining when an AI system can be said to embody certain values. This account understands embodied values as the result of design activities intended to embed those values in such systems. AI systems are here understood as a special kind of sociotechnical system that, like traditional sociotechnical systems, are composed of technical artifacts, human agents, and institutions but—in addition—contain artificial agents and certain technical norms that regulate interactions between artificial agents and other elements of the system. The specific challenges and opportunities of embedding values in AI systems are discussed, and some lessons for better embedding values in AI systems are drawn.

Other Versions

No versions found

Links

PhilArchive



    Upload a copy of this work     Papers currently archived: 98,219

External links

Setup an account with your affiliations in order to access resources via your University's proxy server

Through your library

Analytics

Added to PP
2020-09-02

Downloads
175 (#122,166)

6 months
28 (#116,372)

Historical graph of downloads
How can I increase my downloads?

Author's Profile

References found in this work

We have never been modern.Bruno Latour - 1993 - Cambridge: Harvard University Press.
What We Owe to Each Other.Thomas Scanlon - 2002 - Mind 111 (442):323-354.
Value in ethics and economics.Elizabeth Anderson - 1993 - Cambridge: Harvard University Press.
Minds, Brains and Science.John R. Searle - 1984 - Cambridge: Harvard University Press.

View all 45 references / Add more references