Citations of:
On malfunctioning software
Synthese 192 (4):1199-1220 (2015)
Add citations
You must login to add citations.
|
|
The reviewers Rapaport, Stephanou, Angius, Primiero, and Bringsjord of Turner cover a broad range of topics in the philosophy of computer science. They either challenge the positions outlined in Turner or offer a more refined analysis. This article is a response to their challenges. No categories |
|
The reviewers Rapaport, Stephanou, Angius, Primiero, and Bringsjord of Turner cover a broad range of topics in the philosophy of computer science. They either challenge the positions outlined in Turner or offer a more refined analysis. This article is a response to their challenges. No categories |
|
In this paper, I offer a contribution to the debate on testimony that rests on three elements: the definition of semantic information, the analysis of trust as a second?order property of first?order relations, and Floridi?s Network Theory of Account (NTA). I argue that testimony transmits semantic information and it is neither grounded on trust nor justified by it. Instead, I show that testimony is an occurrence of a first?order relation of communication affected by the second?order property of trust. I then (...) |
|
How can we be certain that software is reliable? Is there any method that can verify the correctness of software for all cases of interest? Computer scientists and software engineers have informally assumed that there is no fully general solution to the verification problem. In this paper, we survey approaches to the problem of software verification and offer a new proof for why there can be no general solution. |
|
How can we be certain that software is reliable? Is there any method that can verify the correctness of software for all cases of interest? Computer scientists and software engineers have informally assumed that there is no fully general solution to the verification problem. In this paper, we survey approaches to the problem of software verification and offer a new proof for why there can be no general solution. |
|
We address some of the epistemological challenges highlighted by the Critical Data Studies literature by reference to some of the key debates in the philosophy of science concerning computational modeling and simulation. We provide a brief overview of these debates focusing particularly on what Paul Humphreys calls epistemic opacity. We argue that debates in Critical Data Studies and philosophy of science have neglected the problem of error management and error detection. This is an especially important feature of the epistemology of (...) No categories |
|
In this paper, I argue that machines can create works of art. My argument is based on an analysis of the so-called creative machines and focuses on technical functions and intentions. If my proposal is correct, then creative machines are technical artifacts with the proper function to bring about works of art. My account is based on sensible conceptual connections between makers, technical artifacts, intentions, and the creation of art. One upshot of the account presented here is that we do (...) |
|
We show how more general knowledge can be built in information security, by the building of knowledge of mechanism clusters, some of which are multifield. By doing this, we address in a novel way the longstanding philosophical problem of how, if at all, we come to have knowledge that is in any way general, when we seem to be confined to particular experiences. We also address the issue of building knowledge of mechanisms by studying an area that is new to (...) No categories |
|
We show how more general knowledge can be built in information security, by the building of knowledge of mechanism clusters, some of which are multifield. By doing this, we address in a novel way the longstanding philosophical problem of how, if at all, we come to have knowledge that is in any way general, when we seem to be confined to particular experiences. We also address the issue of building knowledge of mechanisms by studying an area that is new to (...) No categories |
|
One might poetically muse that computers have the essence both of logic and machines. Through the case of the history of Separation Logic, we explore how this assertion is more than idle poetry. Separation Logic works because it merges the software engineer’s conceptual model of a program’s manipulation of computer memory with the logical model that interprets what sentences in the logic are true, and because it has a proof theory which aids in the crucial problem of scaling the reasoning (...) No categories |
|
Software-intensive science challenges in many ways our current scientific methods. This affects significantly our notion of science and scientific interpretation of the world, driving at the same time the philosophical debate. We consider some issues prompted by SIS in the light of the philosophical categories of ontology and epistemology. |
|
Malware has been around since the 1980s and is a large and expensive security concern today, constantly growing over the past years. As our social, professional and financial lives become more digitalised, they present larger and more profitable targets for malware. The problem of classifying and preventing malware is therefore urgent, and it is complicated by the existence of several specific approaches. In this paper, we use an existing malware taxonomy to formulate a general, language independent functional description of malware (...) No categories |
|
We present a review of Raymond Turner’s Book Computational Artifacts – Towards a Philosophy of Computer Science, focusing on three main topics: Design, Malfunction, and Validity. No categories |
|
The process of completing, correcting and prioritising specifications is an essential but very complex task for the maintenance and improvement of software systems. The preservation of functionalities and the ability to accommodate changes are main objectives of the software development cycle to guarantee system reliability. Logical theories able to fully model such processes are still insufficient. In this paper we propose a full formalisation of such operations on software systems inspired by the Alchourrón–Gärdenfors–Makinson paradigm for belief revision of human epistemic (...) |
|
Machine awareness is a disputed research topic, in some circles considered a crucial step in realising Artificial General Intelligence. Understanding what that is, under which conditions such feature could arise and how it can be controlled is still a matter of speculation. A more concrete object of theoretical analysis is algorithmic iteration for computational intelligence, intended as the theoretical and practical ability of algorithms to design other algorithms for actions aimed at solving well-specified tasks. We know this ability is already (...) |
|
The standard rule of single privative modification replaces privative modifiers by Boolean negation. This rule is valid, for sure, but also simplistic. If an individual a instantiates the privatively modified property (MF) then it is true that a instantiates the property of not being an F, but the rule fails to express the fact that the properties (MF) and F have something in common. We replace Boolean negation by property negation, enabling us to operate on contrary rather than contradictory properties. (...) |
|
The birth, growth, stabilization and subsequent understanding of a new field of practical and theoretical enquiry is always a conceptual process including several typologies of events, phenomena an... |
|
This paper contributes to the computer ethics debate on software ownership protection by examining the ontological, methodological, and ethical problems related to property right infringement that should come prior to any legal discussion. The ontological problem consists in determining precisely what it is for a computer program to be a copy of another one, a largely neglected problem in computer ethics. The methodological problem is defined as the difficulty of deciding whether a given software system is a copy of another (...) |
|
This paper addresses the methodological problem of analysing what it is to explain observed behaviours of engineered computing systems, focusing on the crucial role that abstraction and idealization play in explanations of both correct and incorrect BECS. First, it is argued that an understanding of explanatory requests about observed miscomputations crucially involves reference to the rich background afforded by hierarchies of functional specifications. Second, many explanations concerning incorrect BECS are found to abstract away from descriptions of physical components and processes (...) |
|
In information societies, operations, decisions and choices previously left to humans are increasingly delegated to algorithms, which may advise, if not decide, about how data should be interpreted and what actions should be taken as a result. More and more often, algorithms mediate social processes, business transactions, governmental decisions, and how we perceive, understand, and interact among ourselves and with the environment. Gaps between the design and operation of algorithms and our understanding of their ethical implications can have severe consequences (...) |
|
|
|
Conventional accounts of epistemic opacity, particularly those that stem from the definitive work of Paul Humphreys, typically point to limitations on the part of epistemic agents to account for the distinct ways in which systems, such as computational methods and devices, are opaque. They point, for example, to the lack of technical skill on the part of an agent, the failure to meet standards of best practice, or even the nature of an agent as reasons why epistemically relevant elements of (...) |
|
|