1 Introduction

There is a general or near consensus on the view that the ideal of value-free science is untenable and that science is and ought to be value-laden in some respect (Lusk, 2021; Holman and Wilholt, 2022).Footnote 1 Considerations such as the endemic uncertainty of empirical science, the role of contingency in science, the nature of scientific practice, the pragmatic orientation of scientific inquiry, the way that the public relies on science for advice, and the normative weight of many scientific concepts have figured in a wide variety of arguments against the value-free ideal (VFI). Arguments against the VFI can be traced back at least to the late nineteenth century (James, 1896; Magnus, 2013, 2022) and have grown increasing sophisticated since. The current priorities of the field have largely shifted from arguments about the VFI to questions about how to understand science and its role in society in the face of its value-ladenness.

However, the consensus is not without its challengers; indeed, defenses of the VFI appear with some regularity. The last time there was a concerted response to such challenges by philosophers of science working on values in science was in response to Betz (2013) (see Miller, 2014, John, 2015a, Steel, 2016, Douglas, 2017, Resnik, 2017, Frank, 2017, Lusk, 2021). Betz’s concerns have been answered in many ways, as have many of the arguments that came before. Many of the more recent challenges have not received the same response from the field. Thus, I will focus attention on defenses of the VFI published after 2013, surveying and evaluating them in light of the best arguments against the VFI in the recent literature. I will also consider critiques of arguments against the VFI or for values in science that do not go so far as to defend the VFI.

The literature on science and values at present is large and complex, but two kinds of questions are central. The first question is: should science be value-free, i.e., ought scientists to regulate their practice by aiming at value-freedom? There are various ways one might interpret the value-free ideal (VFI), but the most common way is the claim that only scientific or “epistemic” values can influence scientific reasoning or inference, while the only place for other values, including social and ethical values, should be in external aspects of science, such as choice of research projects or decisions about acceptable methods. I will call this question, about whether the VFI is a correct normative ideal for science, “the VFI question.” The second question is, assuming that science is not value-free, how ought the role of values in science be managed, that is, when and how should values be permitted to operate in science? I will call this “the value-management question.”.Footnote 2

The two questions are often conflated, giving rise to various confusions. One of these confusions is that there are a spectrum of positions on values in science, with the VFI at one extreme, and some sort of radical value-ladenness at the other. This characterization is a mistake, however, given that the VFI is an all-or-nothing affair—either social and ethical values should play a role in the internal phases of scientific reasoning, or they should not. And this matches what most defenders of the VFI intend; they are defending an all-or-nothing claim: no non-epistemic values allowed in the internal phases of science. While different interpretations of key terms (e.g., “epistemic values,” “internal phases”) can lead to somewhat different interpretations or versions of the VFI, any supposed “middle position” on the question is a rejection of the VFI. A related confusion is to treat various positions on the second question as if they were partial defenses of the VFI. I demonstrate this point below, when I discuss the supposed democratic defense of the VFI (which is more productively considered as a constraint on answers to the value-management question) and the various attempts to effect a partial rapprochement between defenses and criticisms of the VFI. Some work that claims to be addressing the VFI question is better understood as a specific sort of answer to the value-management question. It only makes sense to address the latter question once the VFI has been rejected.

This paper takes up the VFI question. My argument is that the best recent criticisms of the VFI set a high bar for defending the VFI, and that recent defenses of the VFI have by and large failed to adequately respond to these criticisms. Many of these defenses raise legitimate concerns against the older arguments that they target, but because they do not address recent moves, they are not successful in their goals. Other arguments raise important concerns about the role of values in science that must be addressed when we take on the value-management question, but they do not succeed as arguments in favor of the VFI.

I will begin with a survey of the strongest arguments against the VFI and the crucial moves and counter-moves that must be taken into account by anyone aiming to defend the VFI. Central emphasis here belongs on the argument from inductive risk as developed by Heather Douglas (2000). I will synthesize this survey into a statement of the major considerations for an adequate defense of the VFI, given the state of the literature. Next, I survey recent defenses of the VFI, grouping related defenses together, and showing that they all fail to meet the challenge posed by the arguments against the VFI. Then, I consider various attempts to affect a partial rapprochement with the VFI, arguing that what is valuable in these arguments is better seen as addressing the value-management question. In conclusion, I argue that we either need better arguments in defense of the VFI, or we need to move past the VFI question and focus our attention on the value-management question and other issues that arise from the rejection of the VFI.

2 Arguments against the VFI

The ideal of value-free science asserts that there is a part of science (sometimes called the “internal” or “inferential” part) where certain values (“non-epistemic values”) ought not be permitted to have an influence (Douglas, 2009, Ch 3; 2016). According to the VFI, only epistemic values are allowed in the context of justification; for this reason, we could call it “the ideal of epistemic purity” instead (Biddle, 2013), though I will stick to the standard terminology. The VFI is typically understood by its defenders to be limited to certain parts of science; many though not all will acknowledge that values have a legitimate role to play in determining which questions or problems scientists pursue or how they frame hypotheses for consideration. It is also widely acknowledged that scientific methods must be constrained by ethical considerations such as the wellbeing of human research subjects or environmental impacts. However, when it comes to the constitution of evidence or evaluating the way that evidence bears in support or against a hypothesis, only epistemic values may play a role. Scientists ought not consider non-epistemic values when making judgments in this part of science. Though there is room to debate which values must be proscribed from which phases of scientific inquiry to achieve the “epistemic purity” that the VFI demands,Footnote 3 to defend the VFI is to insist that science ought to be absolutely value-free within those bounds.

The VFI is a normative ideal, not a descriptive claim. It is meant to govern the kinds of considerations scientists should weigh in making and justifying their decisions, actions, and claims, as well as the norms governing social conventions and institutional structures that mediate the assessment and acceptance of scientific claims. Defenders of the VFI can and do acknowledge that values sometimes cause scientists to act or reason in certain ways, or that values often motivate the way that they pursue their work. Defenders of the VFI readily admit that scientists are human and not perfect epistemic machines. What matters is that scientists explicitly strive for neutrality or impartiality in their explicit reasoning, as well as adopting individual and social practices that tend to minimize the role of nonepistemic values as (direct or indirect) justifying reasons.Footnote 4 As an ideal, the VFI is insulated from crass ought-implies-can arguments. It does not matter whether it is achievable; it may still be worth pursuing (see Menon and Stegenga, 2023). The question that concerns us here is whether scientists in practice ought to be guided by the VFI, whether they ought to take it as a regulative ideal of their activity, or whether they ought to strive towards value-freedom in some sense.

An adequate critique of the VFI must be a normative argument that, all things considered, the VFI is not desirable to pursue or achieve in principle, even if it were possible in practice. Such critiques must show that it is normatively legitimate, even required, for scientists to consider what are typically termed “non-epistemic values” in the ordinary course of their work, in the internal part of science. Recent attacks on the VFI meet this challenge.Footnote 5 In the rest of this section, I will review the strongest of these arguments, and some of the core moves that have been established in the back-and-forth over the VFI.

2.1 The Argument from Inductive Risk (AIR)

The most influential and arguably the strongest argument against the VFI in the recent literature is the argument from inductive risk (AIR). The contemporary form of the AIR is due to the groundbreaking work of Heather Douglas (2000, 2009). Havstad (2022) shows that, although Douglas adopts the term “inductive risk” from Hempel (1965), and despite the fact that Douglas’s argument bears resemblance to earlier arguments from Churchman (1948), Rudner (1953), and others, Douglas’s argument is distinct (296n6, 309n37). Importantly, as Havstad shows, Rudner’s argument is weaker than Douglas’s.

Brown and Stegenga (2023), building on Havstad (2022), have provided a reconstruction of Douglas’s (2000) AIR that shows that the argument is valid and, arguably, sound:

  1. 1.

    If it is not the case that scientists ought to consider the predictable consequences of error (or inductive risk), then it is not the case that scientists are responsible for their actions as scientists.

  2. 2.

    If it is not the case that scientists are responsible for their actions as scientists, then it is not the case that scientists have the same moral responsibilities as the rest of us.

  3. 3.

    Scientists have the same moral responsibilities as the rest of us.

  4. 4.

    Therefore, it is not the case that scientists are not responsible for their actions as scientists.

  5. 5.

    Therefore, it is the case that scientists ought to consider the predictable consequences of error (or inductive risk).

  6. 6.

    Where scientists ought to consider inductive risks and the weighing of inductive risk requires the consideration of non-epistemic consequences, non-epistemic values have a legitimate role to play in the internal stages of science.

  7. 7.

    In the cases discussed by Douglas, the consequences of the choices include clear non-epistemic consequences.

  8. 8.

    So in these cases, scientists should weigh the inductive risks, and doing so requires consideration of clear non-epistemic consequences. ————————

  9. 9.

    Therefore, in the discussed cases, non-epistemic values have a legitimate role to play in the internal stages of science.Footnote 6

This argument is deductively valid. The first part of the argument (1–5) proceeds by two applications of modus tollens. The second part of the argument (5–9) is valid by modus ponens plus conjunction introduction. The conclusion (9) amounts to a denial of the VFI. It is important to point out that the argument applies in those cases in science in which “non-epistemic consequences of error can be foreseen” (Douglas, 2000, 578). Where there are no non-epistemic consequences of error, or those consequences are unforeseeable, there may be no role for non-epistemic values on (this version of) the argument from inductive risk in those specific cases. On the other hand, when there are foreseeable non-epistemic consequences, weighing inductive risks means both considering the uncertainties involved in inference and evaluating the different possible consequences of error.Footnote 7

Building on Douglas (2000; 2003; 2009), Havstad (2022) also ably defends the truth of each of the premises (305–309). What we have here is thus a putatively valid and sound argument to the conclusion that non-epistemic values legitimately act as justifying reasons in science, specifically in justifying how inductive risks are weighed. Because weighing inductive risks is part of the internal aspect of scientific inference, this is an argument against the VFI. As such, any defense of the VFI must show either that, appearances aside, the argument form is invalid, or that one of the premises is false.

As Havstad shows, it would seem that the only really plausible premise to attack is (6). And indeed, historically, many defenses of the VFI can reasonably be reconstructed as attacks on premise (6), attempting to show that scientists need not consider non-epistemic values when confronted with inductive risks that have non-epistemic consequences. Havstad makes two major points in defense of (6). First, it is part of the ordinary practice of science to weigh inductive risks and demand stricter evidence when the consequences of error seem more significant. She draws on Douglas (2021) in considering alterative responses to inductive risk, such as flipping a coin or using only epistemic values to decide the case, arguing that these responses are self-undermining, because these non-standard procedures are motivated by nonepistemic values (Havstad, 2022, 308). Second, Havstad builds on Havstad and Brown (2017) in arguing that responding to inductive risk by deferring or hedging rather than making nonepistemic value judgments is unworkable.

2.2 The deferred decision or hedging response

This second line of attack on premise (6) is a classic response to both Rudner’s argument and the AIR that has been called the “deferred decision response” or the “hedging response.” Examples of this response have been given by Jeffrey (1956), Mitchell (2004),Footnote 8 Betz (2013), and Edenhofer and Kowarsch (2015). These responses argue that, though it is necessary to weigh inductive risks in order to reach conclusions about hypothesis acceptance or rejection, it is not an essential part of scientific work to reach such conclusions. Instead, scientists can defer decisions about acceptance to the context of application, providing instead all the relevant information (such as probabilities of the hypothesis given the evidence) that they would use to make such decisions. The conclusions that scientists do (or should) reach (or, at least, assert in public) are suitably hedged in such a way as to putatively avoid inductive risks entirely. Thus, premise (6) is rejected.

Several quite convincing lines of response have been pursued to this line of argument. The first style of response points towards a certain kind of regress of inductive risks, making it clear that scientists cannot in fact defer or hedge effectively and responsibly. This regress goes in two different directions, concerning whether the regress is downstream or upstream of hypothesis evaluation.

What I will call “the downstream regress of inductive risks” is pointed out already by Rudner (1953). Rudner anticipated the deferred decision response and argues in response that inductive risks are also present in whatever information is presented to the decision-makers to whom value judgments are deferred. If instead of asserting H, the scientists assert \(P(H)=p\), there are inductive risks here as well. Steel (2016) calls this “second-order uncertainty.” The downstream regress has quantitative and qualitative aspects. Quantitatively, the assertion that \(P(H)=p\) is not itself completely certain, but itself has inductive risks associated. Likewise, \(P(P(H)=p)=p'\) and so on. Though the practical significance of these risks may decrease as one travels down the regress, certainty is never reached, and so inductive risk never goes away entirely. At a practical level, whether there are foreseeable non-epistemic consequences associated with decisions about N-order uncertainties cannot be determined in advance, a priori; it is context-specific and itself requires a value judgment. Qualitatively, there are multiple options at play for models and methods of estimating probabilities that lead to different probability ascriptions and so different conclusions. In a related but more technical way, Steel (2015) shows that a Bayesian analysis of confirmation remains subject to the AIR.

There is also an upstream regress of inductive risks. The phenomenon of inductive risk does not only apply to the final judgment about whether the evidence supports a hypothesis or theory. There are many intermediate judgments made in scientific inquiry (or many premises and intermediate inferences made in scientific arguments) that themselves are uncertain, with potentially significant non-epistemic consequences for how they are made. Havstad and Brown (2017) follow Douglas (2009) and Winsberg (2012) in arguing that these upstream decisions are too many, and too complex, to be dealt with through the deferred decision response. What’s more, the consequences of those upstream decisions may not align with those of the downstream decisions, but may bring in unique considerations. One might defer a small subset of value-laden decisions, but not all.

In another line of response, Frisch (2020) shows that the decisions cannot be legitimately deferred or hedged based on a principle articulated by Elliott (2011): the no-passing-the-buck principle. Elliott rightly points out that it can be harmful for scientists to withhold judgment on matters when their technical judgment can help inform decision-makers. As Frisch argues, “One way in which scientists may violate the no-passing-the-buck principle is to commit only to a strongly hedged variant of a hypothesis P while withholding judgment on P itself” (Frisch, 2020, 983). There is a kind of trade-off between informativeness and certainty, which is itself a kind of epistemic risk different from but relevant to inductive risks.Footnote 9 Hedging gets you certainty at the cost of informativeness, but science must be informative and policy-relevant in order to be consequential to decision-makers and thus to fulfill the social role that gives it social and political weight (Elliott, 2011; Steele, 2012; Steel, 2016; Brown, 2018a; Frisch, 2020; see Menon and Stegenga, 2023). It is not an either-or proposition (hedged or informative), but rather a trade-off that must be determined by relevant nonepistemic values. Hedging allows one to mitigate inductive risks only at the cost of limiting the social value of informative science. No one ought to deny that there are contexts where hedging or deferring to some extent turns out to be best; what opponents of the VFI argue is that non-epistemic consequences must be considered in determining whether and how much to hedge, and how much inductive risk must be embraced. They rightly deny that it can be known prior to the specific inquiry that it is always best in every context for scientists to present the most strongly hedged claims. Finding the right balance depends in part on nonepistemic values, thus undermining the argument against premise (6).

These strategies for breaking the implication between non-epistemic consequences of inductive risks and the legitimate consideration of values in science thus fail; such critiques of the AIR are unworkable. The attempt to remove values from playing a justifying role in science has introduced a justifying role for values in science. Thus, it is reasonable to tentatively conclude that the AIR is a sound argument against the VFI.

The AIR is by no means the only important argument against the VFI in the contemporary literature. Another major argument against the VFI concerns the value-laden content of certain kinds of concepts and claims used in many of the sciences, especially in the biological, human, and social sciences (Putnam, 2002; Dupré, 2007; Alexandrova, 2017, 2018; Alexandrova and Fabian, 2022). Biddle and Kukla (2017) and Brown (2020) present even more general arguments concerning the variety of contingencies or epistemic risks beyond inductive risks that require value judgments. These arguments are, however, not only lack a deductively valid and sound presentation; they have also received less attention from defenders of the VFI.

2.3 The VFI and the actual practice of science

Answers to the question of whether science ought to be value-free, whether scientists ought to pursue the VFI, affirmative or negative, concern ideals. They are normative claims about what scientists ought to do or aim for, what science ought to be like. Defenders of the VFI acknowledge that scientific practice often fails to live up to the VFI, but this is no objection to the worth of the VFI. Contemporary opponents of the VFI argue instead that the VFI is unworthy as an ideal, that achieving or even pursuing it is undesirable.

I review the centrality of normative ideals in science to distinguish it from idealized images of science. Idealization of course may play as an important role in philosophy of science as it does in science itself. Perhaps it is useful, in constructing normative ideals, to abstract away from certain realities of scientific practice, and consider somewhat idealized accounts of what science is about. Some defenses of the VFI emphasize the conditions of rational belief. Others emphasize the logic of evidential support. Such defenses attempt to argue that the influence of nonepistemic values in science is irrational or misunderstands the logic of scientific inference. They present abstract accounts of inference or justification to do so. Are these idealized accounts legitimate ways of answering the VFI question?

Although we have already discussed Jeffrey’s defense of the VFI in the context of the deferred decision response, another way to understand his argument is as a claim that scientific inference properly understood does not involve accepting or rejecting hypotheses, but appropriately apportioning credences to hypotheses in light of evidence (in the Bayesian way).Footnote 10 Similarly, Lacey has long argued that the ultimate goal of science is to accumulate a stock of established scientific knowledge, which is the sort of knowledge contained in textbooks, and the decisions the scientific community makes about what counts about such knowledge ought to be value-free (Lacey, 1999). According to Lacey, scientists “hold” a claim in this sense only after the elimination of all legitimate doubt, for all practical purposes, in the extreme long run (Lacey, 2015). As such, there are no inductive risks involved, no contingencies remaining.

A serious problem with this family of argumentative strategies is that the operative ideas of belief, inference, or the stock of scientific knowledge are philosophical abstractions with little purchase on the socially relevant parts of actual scientific practice. The questions that guide this paper are not: Is there some context where it makes sense for philosophers to consider the abstract possibility of value-free science? Is some post hoc rational reconstruction of science possible where nonepistemic values are eliminated? Rather, the questions are: Should the VFI guide scientific practice? Should scientists strive to be value-free? Any normative work idealizes from current reality, and particularly when we are concerned with normative ideals. Otherwise, we would simply be left with descriptions of current practices. The question is, can a normative ideal be taken as a regulative ideal for actual scientific practices? In many cases, philosophers’ description of “science” is such an abstraction that the argument simply has no purchase on questions concerning what scientists can or should do.Footnote 11 Philippi (2020) discusses normative ideals for science in detail, showing that any such ideal (including Kitcher’s value-laden “well-ordered science” ideal) ought still be able to guide action; idealizations introduced in defense of the VFI that rob the VFI of the ability to guide action are thus detrimental to its defense.

Elliott argues that, even if a “non-behavioral” account of belief which is value-free is defensible,Footnote 12 another central (arguably more central) cognitive attitude in science is acceptance rather than belief, where acceptance is understood as using a claim as a premise in reasoning or decisions about how to act (Elliott and Willmes, 2013; Elliott, 2013). There’s no doubt that many acceptance decisions require us to consider value judgments, as Lacey acknowledges (Lacey, 2015). But at the cutting edge of science, and where it is pressing for scientists to deliver information for policy purposes, it is issues of accepting rather than believing or holding a claim that are relevant. One might argue that acceptance decisions should fall to policymakers or consumers of scientific information rather than scientists; but as we have already seen, acceptance decisions cannot universally be deferred in this way. These decisions are a proper part of science, and as they are subject to inductive risks, the AIR shows us that we must make value judgments in the process. (The role of non-scientist stakeholders in these processes should of course not be ignored, but this is a value-management question.)

Science is a social practice; the results of particular scientific inquiries circulate in the community as public assertions, not personal beliefs nor even personal decisions to accept certain claims (Franco, 2017; Brown, 2021). If the purpose of understanding whether or not science is value-free is to provide some form of normative guidance to scientists and not just to explore philosophical abstractions, then defenses of the VFI need to speak to what scientists do and how scientific knowledge actually manifests in practice. Assertions, both as items of public record and as things scientists do are necessarily subject to practical reasoning, i.e., value judgment. This aligns with a revised version of the AIR that Douglas (2021) has articulated recently. The virtue of this pragmatic argument from inductive risk is that it switches our view from more abstract and idealized (in the bad sense) discussions of inference relations towards the realities of scientific practice (Brown, 2020, p. 84).

Another aspect of the social nature of scientific practice is that scientific knowledge is not just a matter of individual cognition, just as it is not a matter of merely abstract inferential relations. In understanding what constitutes good scientific practice, or what is required for objective scientific knowledge, it is not sufficient to focus on individual cognition; indeed, what is good or bad for an individual reasoner may be the opposite for an epistemic community. The norms that ought to guide scientific reasoning apply at the level of community structure and process, on this view, not merely at the level of individual reasoning (Longino, 1990; Kitcher, 1990; Solomon, 2001; Peters, 2021), though the latter may also play a crucial role (Holman and Bruner, 2015, Brown, 2020, pp. 17–18).

2.4 Summary of arguments against the VFI

This section has briefly reviewed the main normative arguments against the VFI in the science and values literature. From this review, we can see a few things that are clearly at stake in arguments for and against the VFI. First, the VFI is not a descriptive claim about what scientists do, nor what they can possibly do; nor are the arguments against the VFI claims about what scientists do or can do. It is a normative issue: critiques of VFI find it normatively inadequate or untenable, because, they argue, values are necessary considerations in justifying decisions made in the course of scientific inquiry or inference.

The key lessons from arguments against the VFI are: (A) The AIR as articulated by Douglas (2000) and explicated by Havstad (2022) and Brown and Stegenga (2023) is the argument to beat; it seems clearly valid and likely sound. A convincing defense of the VFI should show that this argument is invalid or that one of the premises is false. (B) Many arguments in defense of the VFI can be reconstructed as attacks on premise (6) of AIR, but those attacks have been carefully rebutted in a way that further discussions must take into account. A defense of the VFI should not simply repeat the deferred decision or hedging response without countering the rebuttals that have already been made. (C) It is also important not to treat the question as one of an abstract logical structure, but one of the social practices of acceptance, assertion, and knowledge-production in science. This does not mean one should not treat the VFI as an ideal, only that the idea should have some application to the actual social practices of science. To this, I would add a fourth lesson (D): controversial assumptions about the nature of value and of the epistemic/non-epistemic value distinction should not be taken for granted without acknowledgment. Many defenders of the VFI, as well as many opponents, assume that (non-epistemic) values are inherently subjective (private, idiosyncratic) and misleading (biasing), but I have forcefully criticized such views (Brown, 2013b; a; 2018b; 2020, Ch. 3). Likewise, the distinction between epistemic and non-epistemic values and the ability of the distinction to support the VFI has been strongly challenged in the literature (Rooney, 1992, 2017; Longino, 1996; Douglas, 2013b).Footnote 13 Defenders may wish to endorse a subjectivist view of values or a strict epistemic/non-epistemic values dichotomy, but cannot assume these to be uncontroversial premises without qualification.

Insofar as we agree that responsiveness to criticism of the very claims one is defending and the type of arguments one is using to defend them is an important standard for philosophical argument, any defender of the VFI should be willing to consider these points.Footnote 14 That said, addressing all of (A-D) is a complex task. One need not expect that every author addresses all of these points. Perhaps an author wishes to focus exclusively on (A), and while this might require some consideration of (B) and (C), one could reasonably put aside (D) for future work. However, one should expect that defenses of the VFI not simply ignore these issues, and that they both acknowledge the problems and suitably qualify the scope of their argument in response. With respect to (D), one might note that one’s argument relies on a fraught and disputed distinction, while putting off for future work providing an adequate account of the distinction that responds to the criticisms of it. However, at some point the defenders of the VFI must take up the issue. If it is continually put off, it becomes a serious concern about the defensibility of the VFI, as it arguably has.

3 Recent defenses of VFI

In this section, I will review a variety of recent defenses of the VFI. Each of the arguments presents itself as an attempt to defend the VFI or refute an argument against the VFI, though in some cases these arguments are merely entertained rather than fully endorsed (as with the arguments by Bright and Lusk). Attempts to affect a partial rapprochement between the VFI and its critics, or to defend a thesis adjacent to the VFI, or to simply raise concerns about value-laden science will be discussed in the next section. I group these defenses into a few types: there are those who would revive the deferred decision or hedging response, those that focus on the role of science in liberal democracy, those that retreat to a kind of ideal theory, and those that distinguish the value of pursuing versus achieving the VFI. I will point to flaws in each of these arguments that are made clear by the discussion above and which show the arguments are unconvincing as defenses of the VFI, though some point to key desiderata and valuable suggestions in answering the value-management question.

3.1 Deferring and hedging revisited

Several recent defenses of the VFI attempt to revive the deferred decision or hedging response (Henschen, 2021; Cassini, 2022; Carrier, 2022; MacGillivray, 2019). Most of these attempts do not engage with past criticisms of similar versions of this response. Henschen (2021) and Cassini (2022) only reconstruct Rudner’s weaker precursor to the AIR, and do not even consider Douglas’s argument in detail, nor the reconstruction by Havstad that seems to show the argument is sound.Footnote 15 Although many responses have been made to this style of argument, none even appear in the bibliography of Henschen’s paper, except references to Douglas.

The core of Henschen’s argument is three-fold: (1) the regress arguments against the deferred decision response fail, because in practice, scientists typically don’t consider higher-order uncertainties (8–11); (2) if scientists accept or reject hypotheses, they need not do so categorically, but only hypothetically (11–12); (3) we can distinguish between using value judgments or pragmatic considerations in decisions to accept/reject or believe/disbelieve a hypothesis, and only pragmatic considerations are unavoidable (16). In that sense, science might not be entirely epistemically pure, but it can nevertheless remain value free. Although Henschen does not explicitly identify a fault with the form or premises of AIR, we can interpret him (and the other defenses in this category, as we saw in § 2.2) as claiming that premise (6) of AIR is false. (Recall that premise (6) says, where weighing inductive risks involves non-epistemic consequences, it is legitimate to use non-epistemic values in the internal stages of science.)

With respect to Henschen’s consideration of the regress problem for the deferred decision response, he is aware of both the upstream and downstream versions. His argument tries to reduce the upstream regress of inductive risks (IR)—which concerns decisions like the weighing of inductive risks in determining evidence prior to considering how the evidence bears on the hypothesis—to the downstream regress of IR—which concerns second-order uncertainty and the IR with accepting probability assignments. This argument fails, because Henschen confuses IR with uncertainty. Henschen writes, “But the problem with Douglas’s suggestion is that the inductive risk that is present at the three lower stages adds up to the total risk of the null hypothesis in accordance with the laws of probability” (2021, p. 7). Uncertainties can be aggregated in the way Henschen describes; but IR involve not only probabilities but outcomes, which are typically qualitatively different things. The appropriate value judgments concerning aggregated uncertainties at the conclusion of inquiry cannot be the same as value judgments concerning decisions made in intermediate stages of the process, which may concern outcomes quite orthogonal to later decisions.Footnote 16

Henschen’s response to the downstream regress issue is also inadequate. Recall, there are both quantitative and qualitative aspects to the problem; he ignores the qualitative aspect of the problem entirely. With respect to the quantitative aspect, Henschen provides a description of common statistical procedure in the sciences, and then asserts that the last probability asserted by that procedure can legitimately be asserted with no further IR. In effect, Henschen implies that scientists can assign perfectly precise probabilities with absolute certainty, without reference to Steel’s (2016) argument that this cannot be done. Henschen might complain that Steel is unhelpfully idealizing scientific practice, since scientists do not in practice consider these higher-order uncertainties. But this won’t do, as it is also not the case that scientists typically follow the hedge-and-defer strategy either. What’s more, Henschen and Cassini entirely fail to consider the trade-off of certainty and informativeness; their recommendations would have science advisors often passing the buck and providing deleteriously uninformative advice to policymakers.

This last point is relevant to Henschen’s distinction between categorically and hypothetically accepting a hypothesis. Henschen argues that even if it is valuable for scientists to actually accept a hypothesis, they should not do so categorically (making bald claims); they should instead assert the hypothesis “only hypothetically” along with qualifications concerning the probability of the hypothesis and a cost-benefit analysis of relying on the hypothesis for a specific course of action being considered. It is not clear why Henschen believes that this hypothetical hypothesis acceptance is somehow value-free. Although it is more hedged than bald, categorical assertion of the hypothesis, such assertions are still made under uncertainties and require decisions about how to trade-off uncertainty and informativeness. This approach could be a candidate answer to the value-management question, but cannot rehabilitate the VFI.

Henschen also revives a related objection from Levi (1960) that relies on a distinction between belief and action (or acceptance). In bringing on board Levi’s objection, Henschen commits to the sort of “non-behavioral” account of belief whose relevance was criticized in §2.3. Henschen also fails to clearly address what distinguishes epistemic values from non-epistemic, and this leads him astray insofar as he draws a poorly explicated distinction between value judgments and “pragmatic considerations.” Henschen writes, “It is only in the case of value judgments that the antecedents refer to valuations of the utility of specific individuals or groups. In the case of conventional or pragmatic reasons, the antecedents make reference to technical goals” (Henschen, 2021, p. 16). But these goals are not purely epistemic, and it seems clear that pragmatic considerations are a subset of value judgments, no less problematic than other nonepistemic values. Henschen also falsely attributes a concession that these considerations “can be regarded as epistemic values” to Staley (2017). Staley only acknowledges that these values can be considered “extrinsic epistemic values” (sensu Steel, 2010) in the right context; but so can moral and political values.Footnote 17

Cassini (2022) relies on a Bayesian account of model assessment to argue for the VFI in the case of simulation models.Footnote 18 Cassini acknowledges that decisions to accept are practical decisions, susceptible to practical reasoning, but like Jeffrey, Cassini denies that these decisions must be made by scientists, and so he falls into the deferred decision response, best understood as attempting to deny premise (6) of AIR. Cassini is a little better at acknowledging the existing literature, but he dismisses the most relevant argument against his own (Steel, 2015) quickly as “too general to be examined here.” He then asserts, without argument (and falsely), that the downstream regress problem can be solved in purely epistemic terms. The only arguments he addresses in detail are those of Rudner, Douglas, and Winsberg (2012; 2018b; a); in doing so, he really only considers the problem of the upstream regress.

In Cassini’s analysis, all of the upstream IR issues can be aggregated into the prior probabilities associated with the hypothesis, the evidence, and the relevant background knowledge.Footnote 19 This conflates IR with uncertainty in ignoring the qualitative difference between (dis)valued outcomes. Also, if values influence the priors in this way, we know that choice of priors influences the posterior probabilities that result. This is actually sufficient to reject the VFI (Steel, 2015). Cassini’s counter to this point depends on the fact that in the long-run, posterior probabilities should converge. This, like Lacey’s account of value-free “holding,” is irrelevant to much of scientific practice, which takes place in the short- to medium-term, not the long-run. Cassini’s strict separation between scientist and science advisor to policy is likewise an abstraction with little practical import, given the mixing of these roles in practice (see Douglas, 2009, p. 82).Footnote 20

In sum, those who would use the deferred decision response in defense of the VFI have not fully come to grips with the previous responses to it, and so have not succeeded in using it to provide an adequate defense of the VFI.

3.2 The democratic defense

Another argument made by Betz (2013) has continued to receive significant play in the literature: that value-laden science is incompatible with the requirements of liberal democracy. This has been called the democratic defense of the VFI or the political legitimacy argument. The argument is that we must uphold the VFI in order for science to play a legitimate role in democratic governance, which we cannot really do without (see Douglas, 2009, 8 & Ch. 2). For instance, Bright (2018) reconstructs W.E.B. Du Bois as arguing that, if scientists are not seen to be following the VFI, they will lose public support and trust, and this will undermine the attempt to use science to forward socially valuable goals.Footnote 21 Likewise, referring to the AIR and the older underdetermination arguments, Lusk (2021) says: “Despite the recent success these arguments have found, they fail to address one of the central historical motivations for adopting the value-free ideal: political legitimacy” (103).Footnote 22 Others make similar claims (Kappel, 2014–05; Kappel and Zahle, 2019; Carrier, 2022).

This is one of the most serious concerns about value-laden science, and in my view very much a live issue. It is useful to look at Lusk’s version of the argumentFootnote 23:

  1. 2.

    [Legitimacy Premise] No set of non-epistemic values should have an undue influence in coercive democratic political decisions.

  2. 3.

    [Infiltration Premise] If non-epistemic values play a role in the empirical justification of political decisions, then those values have an undue influence[...]

  1. C.

    Therefore, it is not the case that non-epistemic values should play a role in empirical justification in democratic decision making. (104)

If this argument is correct, one of the most important background justifications of the AIR, that science has a clear and significant social impact because of its role in policymaking, becomes a reason instead in support of the VFI. However, each premise of this argument is shaky. Lusk (2021) concentrates on the third premise (infiltration), arguing that it is possible for non-epistemic values to have an influence that does not amount to “undue influence.” Also, it seems likely that some non-epistemic values have a legitimate role to play in democratic decision making, e.g., the ethical values constitutive of science, the political values constitutive of liberal democracy, or the values of the majority, which throws doubt on the legitimacy premise.

As arguments to contradictory conclusions, the political legitimacy argument and the AIR cannot both be sound, but it seems clear that the former is on much shakier ground than the latter. As the political legitimacy argument is typically not accompanied by any direct refutation of premises of the AIR, one suspects something has gone wrong. This can be seen in Wilholt’s (2023) response to Du Bois (as reconstructed by Bright (2018)). According to Du Bois, science must aim solely to discover the truth. But, Wilholt argues, the AIR does not deny that the aim of science is to pursue the truth; rather, the AIR shows that, in pursuing the truth, one meets inductive risks that must be managed, that cannot be settled by the aim of truth. Inductive risks concern the trade-off between, as William James puts it, the two “laws” of thought: “Believe truth! Shun error!” (James, 1896). The VFI cannot do the work Du Bois (via Bright) demands of it, to guide scientific investigation in a trustworthy fashion.

In my view (following Lusk), although many of these arguments are presented as defenses of the VFI, they do not really amount to full defenses despite the intentions of the authors. Rather, they present a challenge that accounts of value-laden science need to address. In other words, they conflate the VFI question with the value-management question; their real value is in challenging us to think carefully about the later in the context of the political roles of science. And the literature has risen to the challenge, shifting a great deal of attention in recent discussions to matters of democratic values and the political legitimacy of value-laden science (e.g., Douglas, 2005, 2012, 2013a, Kitcher, 2011, Pinto and Hicks, 2019, Douglas, 2021, Boulicault and Schroeder, 2021, Schroeder, 2021, Lusk, 2021, Alexandrova and Fabian, 2022).

One could (by reading “legitimate” in the AIR as “politically legitimate”) read this argument as a rejection of premise (6). I do not think this is the right interpretation of AIR (where legitimacy should have both an epistemic and ethical aspect as well), nor does it seem to be what is intended by most of these authors. In any case, this tactic would drive the debate back to the deferred decision response. As none of these arguments adds any new reasons to think that the deferred decision response works, this cannot vindicate the democratic defense of the VFI.

3.3 Ideal theory responses

Another approach to defending the VFI is to retreat from the messy world of scientific practice and the science-policy or science-society interface to the realm of epistemic ideals. This sort of retreat automatically raises worries about whether the resulting argument will have any purchase within scientific practice, but we should not dismiss these arguments out of hand. We should ask, instead, can the ideal theory thus devised shed any light or provide any guidance to scientific practice, or our evaluation of it? Unfortunately, it seems that it cannot.

Hudson (2016) curiously separates two parts of the AIR into two separate arguments he calls “the uncertainty argument” (based on Levi’s (1960) reconstruction of Rudner’s argument) and “the moral argument” (based on Douglas (2009)). This divide et impera strategy is problematic, as it separates key premises that work together to make the argument sound.Footnote 24

In his critique of “the uncertainty argument,” Hudson deals with Rudner’s weaker version of the argument, rather than the stronger version presented by Douglas (2000), to which he gives a version of the deferred decision response. He presents the argument in four steps (following Levi, 1960):

  1. (1)

    The scientist qua scientist accepts or rejects hypotheses.

  2. (2)

    No amount of evidence ever completely confirms or disconfirms any (empirical) hypothesis but only renders it more or less probable.

  3. (3)

    As a consequence of (1) and (2), the scientist must decide how high the probability of a hypothesis relative to the evidence must be before he is warranted in accepting it.

  4. (4)

    The decision required in (3) is a function of how important it will be if a mistake is made in accepting or rejecting a hypothesis.

Hudson repeats Jeffrey’s argument against premise (1), though he acknowledges the problems that have been raised with the deferred decision response. He dismisses these concerns with the claim that VFI is an “epistemic ideal” that sets out what is “epistemically preferable.” This does not really address the terms of the argument about what is legitimate in the course of scientific reasoning. That is, rather than address those problems head-on and tell us what scientists could do to defer effectively and responsibly, he simply retreats to ideal theory. But we are not concerned with what is epistemically preferable, but what is preferable all things considered. We don’t want an epistemic ideal, but a scientific ideal, that is, an ideal to guide scientists who have both epistemic and social duties. By refraining from any arguments about what is all-things-considered preferable, Hudson does not argue that scientists should be guided by the VFI in practice.

Hudson’s objection also depends on the so-called “moral argument” being separated and treated as free-floating from the AIR, making (4) seem unmotivated when there are many ways one might resolve the decision about hypothesis acceptance. Hudson’s analysis of the “uncertainty argument” also confuses the AIR with the older “background assumptions” model of the underdetermination argument by emphasizing his premise (3) as a kind of “gap”; though there are different views about the relation of the AIR and the “gap argument” (see ChoGlueck, 2018), the two are not obviously the same, and the version presented by Hudson is not the deductively valid version discussed above. It is crucial to directly address the strongest version of the argument available, which Hudson failed to do by not addressing Douglas’s argument directly. Today, that means the version presented by Havstad (2022) and tweaked by Brown and Stegenga (2023). To his credit, Hudson recognizes the need to provide a compelling account of values and the epistemic/non-epistemic values distinction. However, while Hudson raises some potential concerns about older arguments about this by Longino (1990) and Rooney (1992), he doesn’t actually provide an account himself, nor does he respond to the kinds of considerations raised by more recent arguments, such as Steel (2010), Douglas (2013b), and Rooney (2017).

What about Hudson’s treatment of “the moral argument”? Here he refers directly to Douglas (2009). He presents Douglas’s argument as a hasty generalization from cases of reasonable ethical limitations on research methods (e.g., to protect research subjects) or research topics (e.g., those with potentially catastrophic applications) to the internal stages of science: “Inspired by these sorts of cases, Douglas’ view is that every scientific decision is at the same time a moral choice” (Hudson, 2016, 178). He sees her argument as “largely based on her intuitions” (179) rather than a sound argument. Hudson’s rebuttal of the “moral argument” fails because it treats the argument as if it is totally disconnected from the issue of inductive risk. It makes Douglas’s claim that evidential standards are value-laden seem unmotivated, whereas the inductive risks faced in the course of policy-relevant science are the precise motivators of the argument. These arguments are not intended to work independently. Hudson’s main objections to the moral argument are that the rejection of the VFI will hurt the authority of science (i.e., its trustworthiness or political legitimacy), and further, that the objectivity of science will suffer. These are legitimate concerns about value-laden science, but they should be dealt with as considerations for the value-management question, as Douglas (2009) has already done, and many following her have developed further. Many have argued that the value-ladenness of science is no threat to its objectivity or trustworthiness (Harding, 1995; Hoyningen-Huene, 2023; Hicks and Lobato, 2022).Footnote 25

Sheykh-Rezaee and Bikaraan-Behesht (2023) argue that the VFI is an “epistemic ideal,” similar to Hudson (2016). Their argument amounts to a kind of burden-shifting argument. Rather than attack the premises of the AIR, with which they identify no fault, they attempt to argue that, despite the conclusion that “non-epistemic values have a legitimate role to play in the internal stages of science,” (even a necessary role, they concede), this is somehow not sufficient to defeat the VFI. Sheykh-Rezaee and Bikaraan-Behesht (2023) claim that any argument against the VFI should meet the following conditions:

  1. (1)

    There are non-epistemic values that should be shown to have some role in science.

  2. (2)

    The role the non-epistemic values play in science should be shown to be at the internal stages of science.

  3. (3)

    The role the non-epistemic values play at the internal stages of science should be shown to be necessary for scientific practice.

  4. (4)

    The role the non-epistemic values necessarily play at the internal stages of science should be shown to be constitutive. (146)

The AIR meets conditions (1–3), as they admit. What they introduce is this “constitutivity” condition; although non-epistemic values play a legitimate, even necessary role in scientific reasoning, they do not play a “constitutively necessary” role, and so the VFI is unscathed. “Constitutively necessary” here is defined as “necessary for pursuing some epistemic goal, and not some practical or ethical goal” in an “(imaginary) epistemically ideal situation” (146–147). But the AIR shows that the weighing of inductive risks is a pervasive requirement in any science involving ampliative inferences with foreseeable real-world consequences. It is not only valid in epistemically non-ideal circumstances; the very epistemic goals of science—believe truth, shun error—require us to make these trade-offs, as Wilholt (2023) has shown.

It is true that the use of non-epistemic values to make the trade-offs is based in ethical rather than exclusively epistemic responsibilities, and on this basis, Sheykh Rezaee and Bikaraan-Behesht argue that it is still not “constitutive.” This suggests that condition (4) is simply question-begging, as any seem to be non-constitutive by definition. More importantly, this criterion seems to have no relevance to the kinds of decisions scientists make in practice, because they concede that it is all-things-considered necessary for scientists to consider non-epistemic values in the internal stages of science; even, it seems, under epistemically ideal conditions. An ideal for scientists must be able to guide scientific practice. Labeling parts of scientific practice “non-constitutive” is idle. When considering practically what scientists should do, Sheykh Rezaee and Bikaraan-Behesht trot out the deferred decision response without responding to any of the criticisms of that response.

3.4 Pursuitworthiness of the VFI

Menon and Stegenga (2023) and Stegenga and Menon (2023) provide some of the best attempts at defending the VFI, in terms of responsiveness to the existing literature. They argue, correctly, that the VFI might still be worth pursuing, even if unattainable. That is, its unattainability does not imply that it is not pursuitworthy. This is sometimes acknowledged in the literature,Footnote 26 though opponents of the VFI obviously argue that it still does not turn out to be good to pursue the VFI. Menon and Stegenga also argue, intriguingly, that the VFI might be worth pursuing even if it would be undesirable to achieve as an end state in principle. They cast the VFI in a helpful new way: it aims at the elimination of “bifurcation points,” i.e., decision points where difference in value judgments made in the course of inquiry would lead scientists to infer different conclusions from the same evidence.Footnote 27 According to their interpretation of the VFI, values should not make a “difference to inference” (Stegenga and Menon, 2023).

Understood this way, Menon and Stegenga still do not argue that science can or should be value-free. They acknowledge that the AIR shows that values should play a role in science in some cases. It is tempting to classify this as a kind of partial rapprochement with the VFI rather than a defense per se, but they insist that it is instead a legitimate alternative framing of the VFI, according to which: “Scientists should [typically] act as if science should be value-free” (Menon and Stegenga, 2023, emphasis added).Footnote 28 Between their reframing of the ideal and of the stakes of the debate (away from what ideal is worth achieving, towards what ideal is worth pursuing), they hope to obviate the key arguments against VFI, including AIR. One tempting response here is that they have simply given up the game on the VFI as most philosophers of science understand it, and so no further consideration is necessary, if the VFI question is our target. In explicitly denying the need to do rebut the AIR, the thinking here goes, they concede the main substantive issue. This is too hasty, however, as there is a substantive issue when it comes action-guiding advice to scientists: should scientists make (appropriately warranted or constrained, non-epistemic) value judgments as a routine part of their inquiries, or should they behave as if science ought to be value free, and thus typically deploy strategies that obviate the need to make value judgments?

Menon and Stegenga provide three reasons to think that value-freedom is worth pursuing even if unattainable and undesirable as an end-state: (i) when values influence scientific reasoning, it becomes less truth-apt; (ii) value-free science advising is more democratic; (iii) public trust in science depends on value-freedom. Their argument for (i) can be seen as cherry-picking examples of bad value-laden science while denying that positive examples are probative. It also problematically assumes that values are never truth-apt.Footnote 29 (iii) is not clearly supported by empirical studies on the relation of trust and value-ladenness (Hicks and Lobato, 2022). (ii) & (iii) are also versions of the democratic defense of VFI, which, I have argued, does not provide a compelling defense of the VFI so much as a desideratum for answers to the value-management question; in this case, the advice is to minimize their influence unless doing so runs afoul of more important constraints, such as those, “concerning resource use, research ethics and action-guidance” (Menon and Stegenga, 2023).

The cherry-picking issue is a common strategy in defenses of the VFI. Menon and Stegenga refer to Lysenko’s critique of Mendelian genetics and pre-1970s androcentric primatology. These are unhelpful examples for their argument, though not uncommon ones. What makes the Lysenko case problematic is not the influence of values so much as the backing of the authoritarian Stalinist regime and the brutal repression of dissent; indeed, values have played a valuable role in the critique and revision of simplistic Mendelianism (Levins and Lewontin, 1985; cf. Graham, 2016). The primatology case is of course a favorite in the context of feminist science studies, but many feminist philosophers of science have argued that one cannot simply see the move away from androcentrism in primatology as a move from value-laden to value-free, but rather as a replacement of androcentric with (better) feminist/egalitarian values (Harding, 1986; Longino, 1992). The pretense of value-freedom served to mask the harmful patriarchal values rather than to avoid their influence. It seems clear that the motivating examples they choose do not settle the case.

When it comes to practical advice for scientists, Menon and Stegenga suggest that scientists should typically adopt value-mitigating strategies, i.e., strategies that will help eliminate bifurcation points, except in cases where moral and practical constraints make adopting those strategies undesirable. (In this way, they hope to accomodate Douglas’s argument that value-free science is irresponsible science.) While there may be contexts where value-mitigation may be a good idea, in other cases it is crucial that scientists use (non-epistemic) values to weigh inductive risks, as they admit. More importantly, there is no way to tell ahead of time which kind of case we are in; so, on Menon’s and Stegenga’s own view, scientists will have to continue weighing non-epistemic values throughout inquiry in order to determine whether value-mitigation is permissible or superior to explicit value judgment. Whether to pursue value-mitigating strategies must be judged in each case according to non-epistemic values, effectively undermining the idea that this approach is value-free (even in their as-if sense). This does not hamper the potential value of their approach as an answer to the value-management question instead of the VFI question, which would be a more productive way to frame their work.

Their recommendation of value-mitigation is also susceptible to another self-undermining worry, raised by Havstad: because value mitigation is not a typical canon of scientific procedure, and is justified in part on the basis of non-epistemic values, they introduce just that which they are attempting to remove. They could argue that many of the strategies for value mitigation that they promote (bias reduction, evidence strengthening, deferral, hedging) are indeed common methodological canons in science. There is a danger here of confusing values and biases (neither implies the other, as argued by Douglas and Elliott, 2022). What’s more, these strategies are typically deployed for reasons other than value mitigation. Whether in this case to prefer value mitigating strategies will require a complex non-epistemic value judgment concerning the values that value mitigation promotes (such as democratic legitimacy and public trust) as well as the values that pull in the opposite direction (including those they acknowledge, such as action guidance, research ethics, responsible use of resources). So the self-undermining concern remains, as far as I can see.

Stegenga and Menon (2023) pursue a similar strategy. The novel move in this paper is a focus on scientific consensus. Values in science (understood as bifurcation points that make a difference to what conclusions are inferred from evidence), they argue, impede the achievement of consensus. They defend what they call strong consensus—not only must scientists agree on the conclusion, but they must all endorse the arguments that lead to the conclusion. It is not clear that this argument hits the mark; as Stegenga and Menon themselves point out, there is no necessary connection between values and consensus. Values might play exactly the role the AIR specifies, but not interfere with strong consensus:

For some scientific hypothesis, values could modulate the acceptable false positive and false negative probabilities for all relevant scientists, decision-makers, or people in general in the same way, such that all people either believe or disbelieve that hypothesis. That, in turn, would entail that values would not threaten the potential of a strong consensus about the hypothesis and thus the status of that hypothesis as scientific knowledge. (Stegenga and Menon, 2023, p. 435)

Indeed, on views like those of Kourany (2010), consensus about specific value judgments is at least as desirable as consensus about factual judgments. So whether or not strong consensus is desirable, it does not tell either for or against the VFI.

This points to an ambiguity in the “difference to inference” characterization of the influence of values in science. Values might make a kind of counterfactual difference to inference if adopting different values (or failing to consider values) would lead to different conclusions, while it nevertheless being the case that only one set of values is in fact considered or even defensible for consideration. This won’t do; their argument instead requires the existence of actual (rather than counterfactual) irresolvable value disagreements influencing science. While it is easy to be pessimistic in the face of certain disputes of this kind, hope on this point is at least as attractive a regulative ideal as the VFI. And the relevance of this hope has been acknowledged at least since Rudner (1953), who argued that, “[A] science of ethics is a necessary requirement if science’s progress toward objectivity is to be continuous” (6). In other words, one might address their concern by adopting an ideal of resolving value disputes rather than an ideal that would have us ignore the moral obligations of scientists. What’s more, potentially any consideration that factors in inferential decisions can figure create a difference to inference; values are no more likely that different hypotheses, data analysis techniques, sets of evidence, or epistemic criteria to do so.

Finally, strong consensus is a highly controversial claim; it is both infeasible and undesirable in the views of many. Stegenga and Menon briefly discuss the views of Paul Feyerabend on this point, as a proponent of the value of dissent, but arguably they miss the force of his argument. Not only is dissent productive for science, on his view, consensus is positively detrimental for scientific knowledge, for broadly Millian reasons (we lose our understanding of the ground and meaning of our beliefs without dissent). Solomon (2001) argues persuasively, though on very different grounds, that there is not much epistemic significance to consensus.Footnote 30 If one’s premises are more controversial than the controversial claim one wishes to argue, the argument is unlikely to convince. The advice to pursue “consensus-forming activity in science” instead of using values to weigh inductive risks likewise falls prey to all the same worries as “value-mitigating strategies.”

It is best to think about what Stegenga and Menon are doing as answering the value-mitigation question. Although they present their work as a defense of the VFI, the changed terms of debate and the nature of their argument really puts them in the camp of the accounts discussed below.

4 Partial rapprochements with the VFI

There is a set of arguments that, while not exactly defending the VFI, seek to accommodate what they take to be right in both defenses and critiques of the VFI by specifying value-laden and value-free moments of scientific inquiry, while acknowledging that the final result is thus value-laden in a sense. Defenders of such views imply that outright rejection of the VFI goes too far, while acknowledging that critics of the VFI have a point. However, this rhetoric presupposes a kind of error, the conflation of the VFI question with the value-management question. One can take a quite conservative view on the value-management question (e.g., Steel, 2010), but this does not uphold the VFI.Footnote 31

Carrier (2022) attempts a qualified defense of the VFI, arguing that while nonepistemic values are essential for certain parts of the scientific process, scientists can nonetheless withhold commitment from those values “by elaborating a plurality of policy packages” that hypothetically involve different values. This appears at first to be a variety of the deferred decision response, very close to Mitchell (2004) and Edenhofer and Kowarsch (2015), one that also fails to account for much of the subsequent debate on that response. Carrier limits his defense of VFI to hypothesis assessment, while he acknowledges the role of values in determining questions, concepts, relevance of evidence, which already concedes the case against the VFI, as these elements cannot reasonably be classified as “external” to scientific reasoning proper. In other words, in response to the upstream regress problem, Carrier freely admits that these decisions are value-laden and must be made by the scientists.

What Carrier perhaps fails to acknowledge here is the point from Okruhlik (1994), which is that any value-free decision-procedure, operating on inputs that are value-laden, will reproduce rather than eliminate that value-ladenness (cf. Elliott and McKaughan, 2009). If value judgments determine in part which hypotheses are proposed, how concepts are operationalized, which evidence is considered relevant, etc., then the final results will be different on the basis of those value judgments, even if the final-stage decision were value-neutral or deferred to others. In Okruhlik’s example, if patriarchal values inform all of the hypotheses under consideration, and if sexist biases determine which evidence is collected and how it is framed, then no matter how value-neutral the inference procedure, it will not erase the influence of those values. Nor, if we turn the final decision over to policymakers, even non-sexist, anti-patriarchal ones, will they be able to eliminate that bias. As Okruhlik puts it, “even if we grant that the standards of theory assessment are free of contamination by non-cognitive factors, nonetheless, non-cognitive values may permeate the very content of science... Even granting the transcendence of method, in other words, the scientific product could itself be radically culture-bound [i.e., value-laden]” (pp. 38–39, emphasis in original). To put it more succinctly: values in, values out.Footnote 32

Carrier does acknowledge that Douglas’s argument is stronger than Rudner’s, and like Mitchell, challenges premise (6). He claims that the relevant information can be presented as a menu of policy options with background values specified, in line with Edenhofer and Kowarsch (2015). Carrier provides no response to the concerns raised about this strategy by e.g. Havstad and Brown (2017). However, this is not to say that, under certain conditions, presenting partially hedged information and deferring certain limited decisions to policymakers is not a fruitful approach to science advising; this could be a good answer to the value-management question. It just does nothing to vindicate the VFI, even in part.

Certain attempted rapprochements focus on sequestering values influences into a certain phase of scientific inquiry that seems less problematic. Recent work by Kareem Khalifa and collaborators, for instance, focuses on the role of questions in inquiry, and explicitly aligns itself with the value-free ideal in the sense of, “epistemic considerations being the only rational determinants for accepting or rejecting hypotheses” (DiMarco and Khalifa, 2019, 1022; cf. Khalifa et al., forthcoming). A related approach focuses on the concept of adequacy-for-purpose in model assessment, which supposedly “opens the possibility of effecting a partial rapprochement between critics and proponents of the value-free ideal” (Lusk and Elliott, 2022). Applying the adequacy-for-purpose approach from model assessment more widely, to all forms of scientific assessment, Lusk and Elliott argue that we may be able to recast what looks like value-laden scientific reasoning into value-free assessment of hypotheses about whether something is adequate for some purpose.Footnote 33 Yet another approach of this type is Wendy Parker’s (2024) recent “epistemic projection approach.”

There are a number of problems with the proposed rapprochements with VFI. First, it is not always possible to specify in advance the range of value considerations, so that they can be loaded into the question or purpose that inquiry begins with. The research question that inquiry is trying to solve may be only vague and inchoate at the start of inquiry and not properly settled until inquiry concludes (Brown, 2020, Ch. 1). The relevant range of options and relevant values may likewise only be discovered in the course of inquiry, not settled ahead of time. Now, the defenders of this view might acknowledge this point, but insist that they can retain their view by packing the values into the posing of new questions throughout the course of inquiry or into the revision of the purposes for the sake of which adequacy judgments are made. This raises several concerns. One is Okruhlik’s “values in, values out” problem. Perhaps most pressingly, it raises the worry about whether these accounts are merely a notational variant of the picture presented by the opponents of the VFI, describing a different way that value judgments figure in the internal processes of scientific inquiry, rather than denying that they do so. It is not clear to me that these views differ significantly from Douglas’s (2009) use of the distinction between direct and indirect roles for values to answer the values management question. This is not to say that such alternative analyses of have no value, only that they do not in fact partially vindicate the VFI; rather, they address the value-management question in a particular way.Footnote 34 Furthermore, assessment of answers to questions or of the adequacy-for-purpose of models (hypotheses, etc) themselves are liable to the same kinds of concerns about the regress of upstream and downstream inductive risks as are the assessments of probabilities assigned to hypotheses given evidence as advocated by the hedging response. It is not clear on this basis whether this rather conservative response to the value-management question is workable. In any case, what is clear is that the aspiration for rapprochement with the VFI, to to retain the rhetoric of epistemic purity, is misplaced.

What’s more, not all legitimate influences of values concern the purposes that guide the asking of questions or the proposal of hypotheses or models. Some values act as side constraints, limiting how we pursue inquiry quite independent of our aims and purposes (Brown, 2017, 72–73). As these sort of value judgments operate no matter what purposes the scientists pursue or what questions they seek to answer, they seem to escape the analysis in question. One can sweep all of this under the rug (that is, into the questions or purposes) only by doing violence to actual scientific practice via a problematic form of rational reconstruction.

John (2015b, 2019) problematizes these issues by focusing on contexts of communication in science, especially public communication, addressing the practical context in which these questions matter. But John’s argument is not without its problems. John (2015b) attempts to improve AIR by restricting the consideration of IR to one’s intended audience, with the aim of matching their values. He motivates this on the intuition that scientists cannot be held responsible for how bad actors interpret their work. This is a mistake; we should reject this amendment because the response of bad actors often forms part of the foreseeable consequences of assertion, i.e., foreseeable perlocutionary effects for which the speaker may be responsible (Franco, 2017). Likewise, we should reject the claim that by virtue of the open-endedness of the audience for scientific publications, we cannot anticipate the reaction of the audience. AIR only requires that we consider the foreseeable consequences of our assertions, and some reactions are clearly foreseeable.

John anticipates this objection, and his response makes two points: first, balancing all the foreseeable consequences of our actions is difficult. While this is true in some cases, in others it is manageable, and one cannot determine which case it is without making a value judgment about the specific case. Second, the costs of error are complex, because it may have many downstream effects that are difficult to weigh. True, but again, to different degrees in different cases. In any event, difficult and complex are not the same as impossible or impracticable. Normatively, the consideration of values is needed, and the difficulty of the task concerns how and not whether inductive risks are managed via value judgments. John also argues that, although AIR applies, scientists should resolve IR by adopting universally high epistemic standards. But the reasons he adduces in favor are defeasible, since it the consequences of false negative error is as relevant as false positive error. One can thus antipate the considerations favoring high epistemic standards being outweighed in specific cases, making the role of value judgments unavoidable in principle, even if high epistemic standards are often the best way to manage IR. There is also the question: how high? It is doubtful that there is any one-size-fits-all answer to this question, especially in light of concerns about second-order uncertainty (Steel, 2016).Footnote 35

These attempts at partial rapprochement with the VFI are valuable contributions to the literature but framed in a misleading and unproductive way. It is not that we need to find a middle ground in the debate for and against the VFI, because the VFI is rightly understood as an “ideal of epistemic purity” (Biddle, 2013), and so any “middle ground” is a rejection of the VFI. Rather, these raise important considerations when addressing the value-management question, and in many cases suggest serious frameworks for answering that question (if not entirely unproblematic ones). The urge towards rapprochement should be rejected in favor of more careful analysis of the issue of managing the roles of values in science.

5 Conclusion

One might object: isn’t this all a bit too fussy? There are some arguments for the VFI and against it, and maybe some middle positions, and a lively debate that we should expect to continue on, as all deep philosophical debates do. Have I not put too much emphasis on one particular reconstruction of AIR? Is that not unfair, especially given the recency of that reconstruction? This is to mistake my point. We have, now, a strong contender for an argument against VFI that is not only valid, but sound. It is the state of the art in the field, drawing on the work by Douglas that gave this issue renewed energy and has made it one of the most vibrant areas of research in philosophy of science. To defeat this argument, a definite flaw in terms of either invalid form or false premise should be identified. Sound deductive arguments are the gold standard for good arguments (Cartwright, 2013), and this one is decisive point against the VFI.

Perhaps you are unpersuaded by this, questioning with Peirce the wisdom in relying on a single deductive argument for a substantive conclusion: “Philosophy ought to... trust rather to the multitude and variety of its arguments than to the conclusiveness of any one” (Peirce, 1868, p. 141). But the AIR is only the tip of the iceberg in terms of a variety of arguments pointing in the same direction. It is the strongest thread in a cable of arguments pulling against the VFI, which also includes the conceptual argument (Dupré, 2007; Alexandrova, 2018), the contingency and epistemic risk arguments (Brown, 2020; Biddle and Kukla, 2017), underdetermination arguments (Longino, 2004, 2008), arguments concerning the permeability of justification by decisions made about pursuit (Okruhlik, 1994), and various others. It would take an overwhelming weight to balance this variety of considerations, and, as I have argued, the weight simply is not there in recent defenses of the VFI.

My hope is that this critical review will have two beneficial effects on the field. First, I hope to have shown that much work in this area might benefit from moving on to the value-management question rather than in revisiting the debate about the VFI. While the VFI-question is an explicitly all-or-nothing question, many of the considerations brought by would-be defenders of the VFI would be better incorporated into more local, more contextual, and more balanced discussions of the relative value of value-laden and value-mitigating strategies in scientific inquiry. I am convinced that this is where the most fruitful lines of new research lay and where the ideas of most of the VFI-defenders would find the most fertile ground. Second, I hope to have provided helpful resources to those who hope to find novel defenses or critiques of the VFI, in terms of a review of the key arguments that ought to be addressed on either side of the debate.