In the past few years, scholars have been questioning whether the current approach in data ethics based on the higher level case studies and general principles is effective. In particular, some have been complaining that such an approach to ethics is difficult to be applied and to be taught in the context of data science. In response to these concerns, there have been discussions about how ethics should be “embedded” in the practice of data science, in the sense of showing (...) how ethical issues emerge in small technical choices made by data scientists in their day-to-day activities, and how such an approach can be used to teach data ethics. However, a precise description of how such proposals have to be theoretically conceived and could be operationalized has been lacking. In this article, we propose a full-fledged characterization of ‘embedding’ ethics, and how this can be applied especially to the problem of teaching data science ethics. Using the emerging model of ‘microethics’, we propose a way of teaching daily responsibility in digital activities that is connected to the higher level ethical challenges discussed in digital/data ethics. We ground this microethical approach into a virtue theory framework, by stressing that the goal of a microethics is to foster the cultivation of moral virtues. After delineating this approach of embedding ethics in theoretical detail, this article discusses a concrete example of how such a ‘micro-virtue ethics’ approach could be practically taught to data science students. (shrink)
In the last few years, biologists and computer scientists have claimed that the introduction of data science techniques in molecular biology has changed the characteristics and the aims of typical outputs (i.e. models) of such a discipline. In this paper we will critically examine this claim. First, we identify the received view on models and their aims in molecular biology. Models in molecular biology are mechanistic and explanatory. Next, we identify the scope and aims of data science (machine learning in (...) particular). These lie mainly in the creation of predictive models which performances increase as data set increases. Next, we will identify a tradeoff between predictive and explanatory performances by comparing the features of mechanistic and predictive models. Finally, we show how this a priori analysis of machine learning and mechanistic research applies to actual biological practice. This will be done by analyzing the publications of a consortium—The Cancer Genome Atlas—which stands at the forefront in integrating data science and molecular biology. The result will be that biologists have to deal with the tradeoff between explaining and predicting that we have identified, and hence the explanatory force of the ‘new’ biology is substantially diminished if compared to the ‘old’ biology. However, this aspect also emphasizes the existence of other research goals which make predictive force independent from explanation. (shrink)
Recently, biologists have argued that data - driven biology fosters a new scientific methodology; namely, one that is irreducible to traditional methodologies of molecular biology defined as the discovery strategies elucidated by mechanistic philosophy. Here I show how data - driven studies can be included into the traditional mechanistic approach in two respects. On the one hand, some studies provide eliminative inferential procedures to prioritize and develop mechanistic hypotheses. On the other, different studies play an exploratory role in providing useful (...) generalizations to complement the procedure of prioritization. Overall this paper aims to shed light on the structure of contemporary research in molecular biology. (shrink)
In the past few years, machine learning (ML) tools have been implemented with success in the medical context. However, several practitioners have raised concerns about the lack of transparency—at the algorithmic level—of many of these tools; and solutions from the field of explainable AI (XAI) have been seen as a way to open the ‘black box’ and make the tools more trustworthy. Recently, Alex London has argued that in the medical context we do not need machine learning tools to be (...) interpretable at the algorithmic level to make them trustworthy, as long as they meet some strict empirical desiderata. In this paper, we analyse and develop London’s position. In particular, we make two claims. First, we claim that London’s solution to the problem of trust can potentially address another problem, which is how to evaluate the reliability of ML tools in medicine for regulatory purposes. Second, we claim that to deal with this problem, we need to develop London’s views by shifting the focus from the opacity of algorithmic details to the opacity of the way in which ML tools are trained and built. We claim that to regulate AI tools and evaluate their reliability, agencies need an explanation of how ML tools have been built, which requires documenting and justifying the technical choices that practitioners have made in designing such tools. This is because different algorithmic designs may lead to different outcomes, and to the realization of different purposes. However, given that technical choices underlying algorithmic design are shaped by value-laden considerations, opening the black box of the design process means also making transparent and motivating (technical and ethical) values and preferences behind such choices. Using tools from philosophy of technology and philosophy of science, we elaborate a framework showing how an explanation of the training processes of ML tools in medicine should look like. (shrink)
Machine learning (ML) has been praised as a tool that can advance science and knowledge in radical ways. However, it is not clear exactly how radical are the novelties that ML generates. In this article, I argue that this question can only be answered contextually, because outputs generated by ML have to be evaluated on the basis of the theory of the science to which ML is applied. In particular, I analyze the problem of novelty of ML outputs in the (...) context of molecular biology. In order to do this, I first clarify the nature of the models generated by ML. Next, I distinguish three ways in which a model can be novel (from the weakest to the strongest). Third, I dissect the way ML algorithms work and generate models in molecular biology and genomics. On these bases, I argue that ML is either a tool to identify instances of knowledge already present and codified, or to generate models that are novel in a weak sense. The notable contribution of ML to scientific discovery in the context of biology is that it can aid humans in overcoming potential bias by exploring more systematically the space of possible hypotheses implied by a theory. (shrink)
In the last few years, biologists and computer scientists have claimed that the introduction of data science techniques in molecular biology has changed the characteristics and the aims of typical outputs (i.e. models) of such a discipline. In this paper we will critically examine this claim. First, we identify the received view on models and their aims in molecular biology. Models in molecular biology are mechanistic and explanatory. Next, we identify the scope and aims of data science (machine learning in (...) particular). These lie mainly in the creation of predictive models which performances increase as data set increases. Next, we will identify a tradeoff between predictive and explanatory performances by comparing the features of mechanistic and predictive models. Finally, we show how this a priori analysis of machine learning and mechanistic research applies to actual biological practice. This will be done by analyzing the publications of a consortium—The Cancer Genome Atlas—which stands at the forefront in integrating data science and molecular biology. The result will be that biologists have to deal with the tradeoff between explaining and predicting that we have identified, and hence the explanatory force of the ‘new’ biology is substantially diminished if compared to the ‘old’ biology. However, this aspect also emphasizes the existence of other research goals which make predictive force independent from explanation. (shrink)
In the past few years, the ethical ramifications of AI technologies have been at the center of intense debates. Considerable attention has been devoted to understanding how a morally responsible practice of data science can be promoted and which values have to shape it. In this context, ethics and moral responsibility have been mainly conceptualized as compliance to widely shared principles. However, several scholars have highlighted the limitations of such a principled approach. Drawing from microethics and the virtue theory tradition, (...) in this paper, we formulate a different approach to ethics in data science which is based on a different conception of “being ethical” and, ultimately, of what it means to promote a morally responsible data science. First, we develop the idea that, rather than only compliance, ethical decision-making consists in using certain moral abilities, which are cultivated by practicing and exercising them in the data science process. An aspect of virtue development that we discuss here is moral attention, which is the ability of data scientists to identify the ethical relevance of their own technical decisions in data science activities. Next, by elaborating on the capability approach, we define a technical act as ethically relevant when it impacts one or more of the basic human capabilities of data subjects. Therefore, rather than “applying ethics”, data scientists should cultivate ethics as a form of reflection on how technical choices and ethical impacts shape one another. Finally, we show how this microethical framework concretely works, by dissecting the ethical dimension of the technical procedures involved in data understanding and preparation of electronic health records. (shrink)
In its last round of publications in September 2012, the Encyclopedia Of DNA Elements (ENCODE) assigned a biochemical function to most of the human genome, which was taken up by the media as meaning the end of ‘Junk DNA’. This provoked a heated reaction from evolutionary biologists, who among other things claimed that ENCODE adopted a wrong and much too inclusive notion of function, making its dismissal of junk DNA merely rhetorical. We argue that this criticism rests on misunderstandings concerning (...) the nature of the ENCODE project, the relevant notion of function and the claim that most of our genome is junk. We argue that evolutionary accounts of function presuppose functions as ‘causal roles’, and that selection is but a useful proxy for relevant functions, which might well be unsuitable to biomedical research. Taking a closer look at the discovery process in which ENCODE participates, we argue that ENCODE’s strategy of biochemical signatures successfully identified activities of DNA elements with an eye towards causal roles of interest to biomedical research. We argue that ENCODE’s controversial claim of functionality should be interpreted as saying that 80 % of the genome is engaging in relevant biochemical activities and is very likely to have a causal role in phenomena deemed relevant to biomedical research. Finally, we discuss ambiguities in the meaning of junk DNA and in one of the main arguments raised for its prevalence, and we evaluate the impact of ENCODE’s results on the claim that most of our genome is junk. (shrink)
We argue that mechanistic models elaborated by machine learning cannot be explanatory by discussing the relation between mechanistic models, explanation and the notion of intelligibility of models. We show that the ability of biologists to understand the model that they work with severely constrains their capacity of turning the model into an explanatory model. The more a mechanistic model is complex, the less explanatory it will be. Since machine learning increases its performances when more components are added, then it generates (...) models which are not intelligible, and hence not explanatory. (shrink)
It has been argued that ethical frameworks for data science often fail to foster ethical behavior, and they can be difficult to implement due to their vague and ambiguous nature. In order to overcome these limitations of current ethical frameworks, we propose to integrate the analysis of the connections between technical choices and sociocultural factors into the data science process, and show how these connections have consequences for what data subjects can do, accomplish, and be. Using healthcare as an example, (...) attention to sociocultural conversion factors relevant to health can help in navigating technical choices that require broader considerations of the sociotechnical system, such as metric tradeoffs in model validation, resulting in better ethical and technical choices. This approach promotes awareness of the ethical dimension of technical choices by data scientists and others, and that can foster the cultivation of 'ethical skills' as integral to data science. (shrink)
Molecular biologists exploit information conveyed by mechanistic models for experimental purposes. In this article, I make sense of this aspect of biological practice by developing Keller’s idea of the distinction between ‘models of’ and ‘models for’. ‘Models of (phenomena)’ should be understood as models representing phenomena and are valuable if they explain phenomena. ‘Models for (manipulating phenomena)’ are new types of material manipulations and are important not because of their explanatory force, but because of the interventionist strategies they afford. This (...) is a distinction between aspects of the same model. In molecular biology, models may be treated either as ‘models of’ or as ‘models for’. By analysing the discovery and characterization of restriction–modification systems and their exploitation for DNA cloning and mapping, I identify the differences between treating a model as a ‘model of’ or as a ‘model for’. These lie in the cognitive disposition of the modeller towards the model: a modeller will look at a model as a ‘model of’ if interested in its explanatory force, or as a ‘model for’ if interested in the material manipulations it can possibly afford. (shrink)
In the last decade, Systems Biology has emerged as a conceptual and explanatory alternative to reductionist-based approaches in molecular biology. However, the foundations of this new discipline need to be fleshed out more carefully. In this paper, we claim that a relational ontology is a necessary tool to ground both the conceptual and explanatory aspects of Systems Biology. A relational ontology holds that relations are prior—both conceptually and explanatory—to entities, and that in the biological realm entities are defined primarily by (...) the context they are embedded within—and hence by the web of relations they are part of. (shrink)
In biology—as in other scientific fields—there is a lively opposition between big and small science projects. In this commentary, I try to contextualize this opposition in the field of biomedicine, and I argue that, at least in this context, big science projects should come first.
We claim that in contemporary studies in molecular biology and biomedicine, the nature of ‘manipulation’ and ‘intervention’ has changed. Traditionally, molecular biology and molecular studies in medicine are considered experimental sciences, whereas experiments take the form of material manipulation and intervention. On the contrary “big science” projects in biology focus on the practice of data mining of biological databases. We argue that the practice of data mining is a form of intervention although it does not require material manipulation. We also (...) suggest that material manipulation, although still present in in the practice of data mining, fulfill a different epistemic role. (shrink)
The applications of machine learning and deep learning to the natural sciences has fostered the idea that the automated nature of algorithmic analysis will gradually dispense human beings from scientific work. In this paper, I will show that this view is problematic, at least when ML is applied to biology. In particular, I will claim that ML is not independent of human beings and cannot form the basis of automated science. Computer scientists conceive their work as being a case of (...) Aristotle’s poiesis perfected by techne, which can be reduced to a number of straightforward rules and technical knowledge. I will show a number of concrete cases where at each level of computational analysis, more is required to ML than just poiesis and techne, and that the work of ML practitioners in biology needs also the cultivation of something analogous to phronesis, which cannot be automated. But even if we knew how to frame phronesis into rules, still this virtue is deeply entrenched in our biological constitution, which computers lack. Whether computers can fully perform scientific practice independently of humans is an ill-posed question. (shrink)
The notion of biological function is fraught with difficulties—intrinsically and irremediably so, we argue. The physiological practice of functional ascription originates from a time when organisms were thought to be designed and remained largely unchanged since. In a secularized worldview, this creates a paradox which accounts of functions as selected effect attempt to resolve. This attempt, we argue, misses its target in physiology and it brings problems of its own. Instead, we propose that a better solution to the conundrum of (...) biological functions is to abandon the notion altogether, a prospect not only less daunting than it appears, but arguably the natural continuation of the naturalisation of biology. (shrink)
A recent article by Herzog provides a much-needed integration of ethical and epistemological arguments in favor of explicable AI in medicine. In this short piece, I suggest a way in which its epistemological intuition of XAI as “explanatory interface” can be further developed to delineate the relation between AI tools and scientific research.
In the last decade, robustness has been extensively mentioned and discussed in biology as well as in the philosophy of the life sciences. Nevertheless, from both fields, someone has affirmed that this debate has resulted in more semantic confusion than in semantic clearness. Starting from this claim, we wish to offer a sort of prima facie map of the different usages of the term. In this manner we would intend to predispose a sort of “semantic platform” which could be exploited (...) by those who wish to discuss or simply use it. We do this by starting from a core distinction between the robustness of representations, which is a philosophy of science issue, and the representations of robustness, which instead pertains to science. We illustrate our proposal with examples from biology, physics and mathematics. (shrink)
I diagnose the current debate between epistemological and ontological emergentism as a Kantian antinomy, which has reasonable but irreconcilable thesis and antithesis. Kantian antinomies have recently returned to contemporary philosophy in part through the work of Luciano Floridi, and the method of levels of abstraction. I use a thought experiment concerning a computer simulation to show how to resolve the epistemological/ontological antinomy about emergence. I also use emergentism and simulations in artificial life to illuminate both levels of abstraction and theoretical (...) challenge for building intelligent agents. (shrink)
Making public policy choices based on available scientific evidence is an ideal condition for any policy making. However, the mechanisms governing these scenarios are complex, non-linear, and, alongside the medical-health and epidemiological issues, involve socio-economic, political, communicative, informational, ethical and epistemological aspects. In this article we analyze the role of scientific evidence when implementing political decisions that strictly depend on it, as in the case of the COVID-19 pandemic. In carrying out this analysis, we will focus above all on the (...) Italian case. This, on the one hand, precisely because Italy led the way regarding the containment policies of the pandemic. Secondly, the government's action was immediately criticized in various respects. Some were calling into question not only the cumbersome political mechanisms, but also suggesting a scarce ability to take scientific evidence into account. On other fronts, there are those who have criticized Italy for its blind and uncritical faith in science and for the paternalism of its decisions. This debate therefore offers the possibility of dealing with some aspects concerning scientific results and their implementation at the political level from the point of view of a political philosophy of science, roughly in the spirit suggested by John Dupré. (shrink)