How the machine ‘thinks’: Understanding opacity in machine learning algorithms

Big Data and Society 3 (1):205395171562251 (2016)
  Copy   BIBTEX

Abstract

This article considers the issue of opacity as a problem for socially consequential mechanisms of classification and ranking, such as spam filters, credit card fraud detection, search engines, news trends, market segmentation and advertising, insurance or loan qualification, and credit scoring. These mechanisms of classification all frequently rely on computational algorithms, and in many cases on machine learning algorithms to do this work. In this article, I draw a distinction between three forms of opacity: opacity as intentional corporate or state secrecy, opacity as technical illiteracy, and an opacity that arises from the characteristics of machine learning algorithms and the scale required to apply them usefully. The analysis in this article gets inside the algorithms themselves. I cite existing literatures in computer science, known industry practices, and do some testing and manipulation of code as a form of lightweight code audit. I argue that recognizing the distinct forms of opacity that may be coming into play in a given application is a key to determining which of a variety of technical and non-technical solutions could help to prevent harm.

Links

PhilArchive



    Upload a copy of this work     Papers currently archived: 91,709

External links

Setup an account with your affiliations in order to access resources via your University's proxy server

Through your library

Similar books and articles

Understanding from Machine Learning Models.Emily Sullivan - 2022 - British Journal for the Philosophy of Science 73 (1):109-133.
Democratizing Algorithmic Fairness.Pak-Hang Wong - 2020 - Philosophy and Technology 33 (2):225-244.

Analytics

Added to PP
2020-11-24

Downloads
131 (#139,348)

6 months
14 (#176,812)

Historical graph of downloads
How can I increase my downloads?