This Element presents a philosophical exploration of the concept of the 'model organism' in contemporary biology. Thinking about model organisms enables us to examine how living organisms have been brought into the laboratory and used to gain a better understanding of biology, and to explore the research practices, commitments, and norms underlying this understanding. We contend that model organisms are key components of a distinctive way of doing research. We focus on what makes model organisms an important type of model, (...) and how the use of these models has shaped biological knowledge, including how model organisms represent, how they are used as tools for intervention, and how the representational commitments linked to their use as models affect the research practices associated with them. This title is available as Open Access on Cambridge Core. (shrink)
This paper aims to identify the key characteristics of model organisms that make them a specific type of model within the contemporary life sciences: in particular, we argue that the term “model organism” does not apply to all organisms used for the purposes of experimental research. We explore the differences between experimental and model organisms in terms of their material and epistemic features, and argue that it is essential to distinguish between their representational scope and representational target. We also examine (...) the characteristics of the communities who use these two types of models, including their research goals, disciplinary affiliations, and preferred practices to show how these have contributed to the conceptualization of a model organism. We conclude that model organisms are a specific subgroup of organisms that have been standardized to fit an integrative and comparative mode of research, and that it must be clearly distinguished from the broader class of experimental organisms. In addition, we argue that model organisms are the key components of a unique and distinctively biological way of doing research using models.Keywords: Experimental organism; Genetics; Model organism; Modeling; Philosophy of biology; Representation. (shrink)
This groundbreaking, open access volume analyses and compares data practices across several fields through the analysis of specific cases of data journeys. It brings together leading scholars in the philosophy, history and social studies of science to achieve two goals: tracking the travel of data across different spaces, times and domains of research practice; and documenting how such journeys affect the use of data as evidence and the knowledge being produced. The volume captures the opportunities, challenges and concerns involved in (...) making data move from the sites in which they are originally produced to sites where they can be integrated with other data, analysed and re-used for a variety of purposes. The in-depth study of data journeys provides the necessary ground to examine disciplinary, geographical and historical differences and similarities in data management, processing and interpretation, thus identifying the key conditions of possibility for the widespread data sharing associated with Big and Open Data. The chapters are ordered in sections that broadly correspond to different stages of the journeys of data, from their generation to the legitimisation of their use for specific purposes. Additionally, the preface to the volume provides a variety of alternative “roadmaps” aimed to serve the different interests and entry points of readers; and the introduction provides a substantive overview of what data journeys can teach about the methods and epistemology of research. (shrink)
This paper proposes an account of scientific data that makes sense of recent debates on data-driven and ‘big data’ research, while also building on the history of data production and use particularly within biology. In this view, ‘data’ is a relational category applied to research outputs that are taken, at specific moments of inquiry, to provide evidence for knowledge claims of interest to the researchers involved. They do not have truth-value in and of themselves, nor can they be seen as (...) straightforward representations of given phenomena. Rather, they are fungible objects defined by their portability and prospective usefulness as evidence. (shrink)
In this paper, we analyse the relation between the use of environmental data in contemporary health sciences and related conceptualisations and operationalisations of the notion of environment. We consider three case studies that exemplify a different selection of environmental data and mode of data integration in data-intensive epidemiology. We argue that the diversification of data sources, their increase in scale and scope, and the application of novel analytic tools have brought about three significant conceptual shifts. First, we discuss the EXPOsOMICS (...) project, an attempt to integrate genomic and environmental data which suggests a reframing of the boundaries between external and internal environments. Second, we explore the MEDMI platform, whose efforts to combine health, environmental and climate data instantiate a reframing and expansion of environmental exposure. Third, we illustrate how extracting epidemiological insights from extensive social data collected by the CIDACS institute yields innovative attributions of causal power to environmental factors. Identifying these shifts highlights the benefits and opportunities of new environmental data, as well as the challenges that such tools bring to understanding and fostering health. It also emphasises the constraints that data selection and accessibility pose to scientific imagination, including how researchers frame key concepts in health-related research. (shrink)
Knowledge-making practices in biology are being strongly affected by the availability of data on an unprecedented scale, the insistence on systemic approaches and growing reliance on bioinformatics and digital infrastructures. What role does theory play within data-intensive science, and what does that tell us about scientific theories in general? To answer these questions, I focus on Open Biomedical Ontologies, digital classification tools that have become crucial to sharing results across research contexts in the biological and biomedical sciences, and argue that (...) they constitute an example of classificatory theory. This form of theorizing emerges from classification practices in conjunction with experimental know-how and expresses the knowledge underpinning the analysis and interpretation of data disseminated online. (shrink)
A heated debate surrounds the significance of reproducibility as an indicator for research quality and reliability, with many commentators linking a "crisis of reproducibility" to the rise of fraudulent, careless and unreliable practices of knowledge production. Through the analysis of discourse and practices across research fields, I point out that reproducibility is not only interpreted in different ways, but also serves a variety of epistemic functions depending on the research at hand. Given such variation, I argue that the uncritical pursuit (...) of reproducibility as an overarching epistemic value is misleading and potentially damaging to scientific advancement. Requirements for reproducibility, however they are interpreted, are one of many available means to secure reliable research outcomes. Furthermore, there are cases where the focus on enhancing reproducibility turns out not to foster high-quality research. Scientific communities and Open Science advocates should learn from inferential reasoning from irreproducible data, and promote incentives for all researchers to explicitly and publicly discuss their methodological commitments, the ways in which they learn from mistakes and problems in everyday practice, and the strategies they use to choose which research component of any project needs to be preserved in the long term, and how. (shrink)
I propose a framework that explicates and distinguishes the epistemic roles of data and models within empirical inquiry through consideration of their use in scientific practice. After arguing that Suppes’ characterization of data models falls short in this respect, I discuss a case of data processing within exploratory research in plant phenotyping and use it to highlight the difference between practices aimed to make data usable as evidence and practices aimed to use data to represent a specific phenomenon. I then (...) argue that whether a set of objects functions as data or models does not depend on intrinsic differences in their physical properties, level of abstraction or the degree of human intervention involved in generating them, but rather on their distinctive roles towards identifying and characterizing the targets of investigation. The paper thus proposes a characterization of data models that builds on Suppes’ attention to data practices, without however needing to posit a fixed hierarchy of data and models or a highly exclusionary definition of data models as statistical constructs. (shrink)
I propose a framework that explicates and distinguishes the epistemic roles of data and models within empirical inquiry through consideration of their use in scientific practice. After arguing that Suppes’ characterization of data models falls short in this respect, I discuss a case of data processing within exploratory research in plant phenotyping and use it to highlight the difference between practices aimed to make data usable as evidence and practices aimed to use data to represent a specific phenomenon. I then (...) argue that whether a set of objects functions as data or models does not depend on intrinsic differences in their physical properties, level of abstraction or the degree of human intervention involved in generating them, but rather on their distinctive roles towards identifying and characterizing the targets of investigation. The paper thus proposes a characterization of data models that builds on Suppes’ attention to data practices, without however needing to posit a fixed hierarchy of data and models or a highly exclusionary definition of data models as statistical constructs. (shrink)
This paper discusses what it means and what it takes to integrate data in order to acquire new knowledge about biological entities and processes. Maureen O’Malley and Orkun Soyer have pointed to the scientific work involved in data integration as important and distinct from the work required by other forms of integration, such as methodological and explanatory integration, which have been more successful in captivating the attention of philosophers of science. Here I explore what data integration involves in more detail (...) and with a focus on the role of data-sharing tools, like online databases, in facilitating this process; and I point to the philosophical implications of focusing on data as a unit of analysis. I then analyse three cases of data integration in the field of plant science, each of which highlights a different mode of integration: inter-level integration, which involves data documenting different features of the same species, aims to acquire an interdisciplinary understanding of organisms as complex wholes and is exemplified by research on Arabidopsis thaliana; cross-species integration, which involves data acquired on different species, aims to understand plant biology in all its different manifestations and is exemplified by research on Miscanthus giganteus; and translational integration, which involves data acquired from sources within as well as outside academia, aims at the provision of interventions to improve human health and is exemplified by research on Phytophtora ramorum. Recognising the differences between these efforts sheds light on the dynamics and diverse outcomes of data dissemination and integrative research; and the relations between the social and institutional roles of science, the development of data-sharing infrastructures and the production of scientific knowledge. (shrink)
Community databases have become crucial to the collection, ordering and retrieval of data gathered on model organisms, as well as to the ways in which these data are interpreted and used across a range of research contexts. This paper analyses the impact of community databases on research practices in model organism biology by focusing on the history and current use of four community databases: FlyBase, Mouse Genome Informatics, WormBase and The Arabidopsis Information Resource. We discuss the standards used by the (...) curators of these databases for what counts as reliable evidence, acceptable terminology, appropriate experimental set-ups and adequate materials (e.g., specimens). On the one hand, these choices are informed by the collaborative research ethos characterising most model organism communities. On the other hand, the deployment of these standards in databases reinforces this ethos and gives it concrete and precise instantiations by shaping the skills, practices, values and background knowledge required of the database users. We conclude that the increasing reliance on community databases as vehicles to circulate data is having a major impact on how researchers conduct and communicate their research, which affects how they understand the biology of model organisms and its relation to the biology of other species. (shrink)
Scientific knowledge production is currently affected by the dissemination of data on an unprecedented scale. Technologies for the automated production and sharing of vast amounts of data have changed the way in which data are handled and interpreted in several scientific domains, most notably molecular biology and biomedicine. In these fields, the activity of data gathering has become increasingly technology-driven, with machines such as next generation genome sequencers and mass spectrometers generating billions of data points within hours, and with little (...) need for human supervision. Given the relative ease and low costs with which datasets can be produced (that is, once a laboratory has been able to afford .. (shrink)
Research, innovation, and progress in the life sciences are increasingly contingent on access to large quantities of data. This is one of the key premises behind the “open science” movement and the global calls for fostering the sharing of personal data, datasets, and research results. This paper reports on the outcomes of discussions by the panel “Open science, data sharing and solidarity: who benefits?” held at the 2021 Biennial conference of the International Society for the History, Philosophy, and Social Studies (...) of Biology, and hosted by Cold Spring Harbor Laboratory. (shrink)
Bogen and Woodward characterized data as embedded in the context in which they are produced (‘local’) and claims about phenomena as retaining their significance beyond that context (‘nonlocal’). This view does not fit sciences such as biology, which successfully disseminate data via packaging processes that include appropriate labels, vehicles, and human interventions. These processes enhance the evidential scope of data and ensure that claims about phenomena are understood in the same way across research communities. I conclude that the degree of (...) locality of both data and claims about phenomena varies depending on the packaging used to make them travel and on the research setting in which they are used. †To contact the author, please write to: ESRC Centre for Genomics in Society, University of Exeter, Byrne House, St. Germans Road, EX4 4PJ Exeter, United Kingdom; e‐mail: [email protected] (shrink)
This article considers the temporal dimension of data processing and use and the ways in which it affects the production and interpretation of knowledge claims. I start by distinguishing the time at which data collection, dissemination, and analysis occur from the time in which the phenomena for which data serve as evidence operate. Building on the analysis of two examples of data reuse from modeling and experimental practices in biology, I then argue that Dt affects how researchers select and interpret (...) data as evidence and identify and understand phenomena. (shrink)
What is the best way to analyse abstraction in scientific modelling? I propose to focus on abstracting as an epistemic activity, which is achieved in different ways and for different purposes depending on the actual circumstances of modelling and the features of the models in question. This is in contrast to a more conventional use of the term ‘abstract’ as an attribute of models, which I characterise as black-boxing the ways in which abstraction is performed and to which epistemological advantage. (...) I exemplify my claims through a detailed reconstruction of the practices involved in creating two types of models of the flowering plant Arabidopsisthaliana, currently the best-known model organism in plant biology. This leads me to distinguish between two types of abstraction processes: the ‘material abstracting’ required in the production of Arabidopsis specimens and the ‘intellectual abstracting’ characterising the elaboration of visual models of Arabidopsis genomics. Reflecting on the differences between these types of abstracting helps to pin down the epistemic skills and research commitments used by researchers to produce each model, thus clarifying how models are handled by researchers and with which epistemological implications. (shrink)
Whether we live in a world of autonomous things, or a world of interconnected processes in constant flux, is an ancient philosophical debate. Modern biology provides decisive reasons for embracing the latter view. How does one understand the practices and outputs of science in such a dynamic, ever-changing world - and particularly in an emergency situation such as the COVID-19 pandemic, where scientific knowledge has been regarded as bedrock for decisive social interventions? We argue that key to answering this question (...) is to consider the role of the activity of reification within the research process. Reification consists in the identification of more or less stable features of the flux, and treating these as constituting stable things. As we illustrate with reference to biological and biomedical research on COVID-19, reification is a necessary component of any process of inquiry and comes in at least two forms: means reification, when researchers create objects meant to capture features of the world, or phenomena, in order to be able to study them; and target reification, when researchers infer an understanding of phenomena from an investigation of the epistemic objects created to study them. We note that both objects and phenomena are dynamic processes and argue that have no reason to assume that changes in objects and phenomena track one another. We conclude that failure to acknowledge these forms of reification and their epistemic role in scientific inquiry can have dire consequences for how the resulting knowledge is interpreted and used. (shrink)
This editorial critically engages with the understanding of openness by attending to how notions of presence and absence come bundled together as part of efforts to make open. This is particularly evident in contemporary discourse around data production, dissemination, and use. We highlight how the preoccupations with making data present can be usefully analyzed and understood by tracing the related concerns around what is missing, unavailable, or invisible, which unvaryingly but often implicitly accompany debates about data and openness.
Arabidopsis is currently the most popular and well-researched model organism in plant biology. This paper documents this plant's rise to scientific fame by focusing on two interrelated aspects of Arabidopsis research. One is the extent to which the material features of the plant have constrained research directions and enabled scientific achievements. The other is the crucial role played by the international community of Arabidopsis researchers in making it possible to grow, distribute and use plant specimen that embody these material features. (...) I argue that at least part of the explosive development of this research community is due to its successful standardisation and to the subsequent use of Arabidopsis specimen as material models of plants. I conclude that model organisms have a double identity as both samples of nature and artifacts representing nature. It is the resulting ambivalence in their representational value that makes them attractive research tools for biologists. (shrink)
The use of big data to investigate the spread of infectious diseases or the impact of the built environment on human wellbeing goes beyond the realm of traditional approaches to epidemiology, and includes a large variety of data objects produced by research communities with different methods and goals. This paper addresses the conditions under which researchers link, search and interpret such diverse data by focusing on “data mash-ups”—that is the linking of data from epidemiology, biomedicine, climate and environmental science, which (...) is typically achieved by holding one or more basic parameters, such as geolocation, as invariant. We argue that this strategy works best when epidemiologists interpret localisation procedures through an idiographic perspective that recognises their context-dependence and supports a critical evaluation of the epistemic value of geolocation data whenever they are used for new research purposes. Approaching invariants as strategic constructs can foster data linkage and re-use, and support carefully-targeted predictions in ways that can meaningfully inform public health. At the same time, it explicitly signals the limitations in the scope and applicability of the original datasets incorporated into big data collections, and thus the situated nature of data linkage exercises and their predictive power. (shrink)
Open Science policies encourage researchers to disclose a wide range of outputs from their work, thus codifying openness as a specific set of research practices and guidelines that can be interpreted and applied consistently across disciplines and geographical settings. In this paper, we argue that this “one-size-fits-all” view of openness sidesteps key questions about the forms, implications, and goals of openness for research practice. We propose instead to interpret openness as a dynamic and highly situated mode of valuing the research (...) process and its outputs, which encompasses economic as well as scientific, cultural, political, ethical, and social considerations. This interpretation creates a critical space for moving beyond the economic definitions of value embedded in the contemporary biosciences landscape and Open Science policies, and examining the diversity of interests and commitments that affect research practices in the life sciences. To illustrate these claims, we use three case studies that highlight the challenges surrounding decisions about how––and how best––to make things open. These cases, drawn from ethnographic engagement with Open Science debates and semistructured interviews carried out with UK-based biologists and bioinformaticians between 2013 and 2014, show how the enactment of openness reveals judgments about what constitutes a legitimate intellectual contribution, for whom, and with what implications. (shrink)
The collection and dissemination of data on human and nonhuman organisms has become a central feature of 21st-century biology and has been endorsed by funding agencies in the United States and Europe as crucial to translating biological research into therapeutic and agricultural innovation. Large molecular data sets, often referred to as “big data,” are increasingly incorporated into digital databases, many of which are freely accessible online. These data have come to be seen as resources that play a key role in (...) mediating global market exchange, thus achieving a prominent social and economic status well beyond science itself. At the same time, calls to make all such data publicly and freely available have garnered strength and visibility, most prominently in the form of the Open Data movement. I discuss these developments by considering the conditions under which data journey across the communities and institutions implicated in globalized biology and biomedicine, and what this indicates about how Internet-based communication and the use of online databases affect scientific research and its role within contemporary society. (shrink)
Many biologists appeal to the so-called Krogh principle when justifying their choice of experimental organisms. The principle states that “for a large number of problems there will be some animal of choice, or a few such animals, on which it can be most conveniently studied”. Despite its popularity, the principle is often critiqued for implying unwarranted generalizations from optimal models. We argue that the Krogh principle should be interpreted in relation to the historical and scientific contexts in which it has (...) been developed and used. We interpret the Krogh Principle as a heuristic, i.e., as a recommendation to approach biological problems through organisms where a specific trait or physiological mechanism is expected to be most distinctively displayed or most experimentally accessible. We designate these organisms “Krogh organisms.” We clarify the differences between uses of model organisms and non-standard Krogh organisms. Among these is the use of Krogh organisms as “negative models” in biomedical research, where organisms are chosen for their dissimilarity to human physiology. Importantly, the representational scope of Krogh organisms and the generalizability of their characteristics are not fixed or assumed but explored through experimental studies. Research on Krogh organisms is steeped in the comparative method characteristic of zoology and comparative physiology, in which studies of biological variation produce insights into general physiological constraints. Accordingly, we conclude that the Krogh principle exemplifies the advantages of studying biological variation as a strategy to produce generalizable insights. (shrink)
ArgumentWe examine the criteria used to validate the use of nonhuman organisms in North-American alcohol addiction research from the 1950s to the present day. We argue that this field, where the similarities between behaviors in humans and non-humans are particularly difficult to assess, has addressed questions of model validity by transforming the situatedness of non-human organisms into an experimental tool. We demonstrate that model validity does not hinge on the standardization of one type of organism in isolation, as often the (...) case with genetic model organisms. Rather, organisms are viewed as necessarily situated: they cannot be understood as a model for human behavior in isolation from their environmental conditions. Hence the environment itself is standardized as part of the modeling process; and model validity is assessed with reference to the environmental conditions under which organisms are studied. (shrink)
The use of big data to investigate the spread of infectious diseases or the impact of the built environment on human wellbeing goes beyond the realm of traditional approaches to epidemiology, and includes a large variety of data objects produced by research communities with different methods and goals. This paper addresses the conditions under which researchers link, search and interpret such diverse data by focusing on “data mash-ups”—that is the linking of data from epidemiology, biomedicine, climate and environmental science, which (...) is typically achieved by holding one or more basic parameters, such as geolocation, as invariant. We argue that this strategy works best when epidemiologists interpret localisation procedures through an idiographic perspective that recognises their context-dependence and supports a critical evaluation of the epistemic value of geolocation data whenever they are used for new research purposes. Approaching invariants as strategic constructs can foster data linkage and re-use, and support carefully-targeted predictions in ways that can meaningfully inform public health. At the same time, it explicitly signals the limitations in the scope and applicability of the original datasets incorporated into big data collections, and thus the situated nature of data linkage exercises and their predictive power. (shrink)
Scientific classification has long been recognized as involving a specific style of reasoning and doing research, and as occasionally affecting the development of scientific theories. However, the role played by classificatory activities in generating theories has not been closely investigated within the philosophy of science. I argue that classificatory systems can themselves become a form of theory, which I call classificatory theory, when they come to formalize and express the scientific significance of the elements being classified. This is particularly evident (...) in some of the classification practices used in contemporary experimental biology, such as bio-ontologies used to classify genomic data and typologies used to classify “normal” stages of development in developmental biology. In this paper, I explore some characteristics of classificatory theories and ways in which they differ from other types of scientific theories and other components of scientific epistemology, such as models and background assumptions. (shrink)
This article documents how biomedical researchers in the United Kingdom understand and enact the idea of “openness.” This is of particular interest to researchers and science policy worldwide in view of the recent adoption of pioneering policies on Open Science and Open Access by the U.K. government—policies whose impact on and implications for research practice are in need of urgent evaluation, so as to decide on their eventual implementation elsewhere. This study is based on 22 in-depth interviews with U.K. researchers (...) in systems biology, synthetic biology, and bioinformatics, which were conducted between September 2013 and February 2014. Through an analysis of the interview transcripts, we identify seven core themes that characterize researchers’ understanding of openness in science and nine factors that shape the practice of openness in research. Our findings highlight the implications that Open Science policies can have for research processes and outcomes and provide recommendations for enhancing their content, effectiveness, and implementation. (shrink)
The paper problematises the reliability and ethics of using social media data, such as sourced from Twitter or Instagram, to carry out health-related research. As in many other domains, the opportunity to mine social media for information has been hailed as transformative for research on well-being and disease. Considerations around the fairness, responsibilities and accountabilities relating to using such data have often been set aside, on the understanding that as long as data were anonymised, no real ethical or scientific issue (...) would arise. We first counter this perception by emphasising that the use of social media data in health research can yield problematic and unethical results. We then provide a conceptualisation of methodological data fairness that can complement data management principles such as FAIR by enhancing the actionability of social media data for future research. We highlight the forms that methodological data fairness can take at different stages of the research process and identify practical steps through which researchers can ensure that their practices and outcomes are scientifically sound as well as fair to society at large. We conclude that making research data fair as well as FAIR is inextricably linked to concerns around the adequacy of data practices. The failure to act on those concerns raises serious ethical, methodological and epistemic issues with the knowledge and evidence that are being produced. (shrink)
: I argue that Open Science as currently conceptualised and implemented does not take sufficient account of epistemic diversity within research. I use three case studies to exemplify how Open Science threatens to privilege some forms of inquiry over others, thus exasperating divides within and across systems of practice, and overlooking important sources and forms of epistemic diversity. Building on insights from pluralist philosophy, I then identify four aspects of diverse research practices that should serve as reference points for debates (...) around Open Science: specificity to local conditions, entrenchment within repertoires, permeability to newcomers and demarcation strategies. (shrink)
We continue our discussion of the competing arguments in favour of the unified theory and the pluralistic theory of radiation advanced by three nineteenth-century pioneers: Herschel, Melloni, and Draper. Our narrative is structured by a consideration of the epistemic criteria relevant to theory-choice; the epistemic focus highlights many little-known aspects of this relatively well-known episode. We argue that the acceptance of light-heat unity in this period cannot be credibly justified on the basis of common evaluative criteria such as simplicity and (...) standard notions of explanatory power. Whether the consensus was justified by some other criteria remains an open question.Keywords: Theory-choice; Epistemic values; Radiation; Simplicity; Explanatory power; Causal explanation. (shrink)
Hardly any ontological result of modern science is more firmly established than the fact that infrared radiation differs from light only in wavelength; this is part of the modern conception of the continuous spectrum of electromagnetic radiation reaching from radio waves to gamma radiation. Yet, like many such evident truths, the light-infrared unity was an extremely difficult thing to establish. We examine the competing arguments in favour of the unified and pluralistic theories of radiation, as put forward in the first (...) half of the nineteenth century by three of the most important early pioneers of the study of radiation: Herschel, Melloni and Draper. In this part of the paper, we conclude that there were no compelling reasons of observational adequacy to prefer the unified theory to the pluralistic theory. Keywords: Macedonio Melloni; John William Draper; William Herschel; Infrared; Observation; Theory-choice. (shrink)
Karen-Sue Taussig: Ordinary Genomes: Science, Citizenship and Genetic Identities Content Type Journal Article Category Book Review Pages 1-4 DOI 10.1007/s10441-012-9150-8 Authors Sabina Leonelli, Department of Sociology and Philosophy, ESRC Centre for Genomics in Society, University of Exeter, Exeter, Devon, UK Journal Acta Biotheoretica Online ISSN 1572-8358 Print ISSN 0001-5342.