Jason Stanley and Timothy Williamson's influential article "Knowing How" argues that knowledge-how is a species of knowledge-that. One objection to their view is that knowledge-how is significantly different than knowledge-that because Gettier cases afflict the latter but not the former. Stanley and Williamson argue that this objection fails. Their response, however, is not adequate. Moreover, I sketch a plausible argument that knowledge-how is not susceptible to Gettier cases. This suggests a significant distinction between knowledge-that and knowledge-how.
Intellectualism about knowledge-how is the view that practical knowledge is a species of propositional knowledge. I argue that this view is undermined by a difference in properties between knowledge-how and both knowledge-that and knowledge-wh. More specifically, I argue that both knowledge-that and knowledge-wh are easily transmitted via testimony while knowledge-how is not easily transmitted by testimony. This points to a crucial difference in states of knowledge. I also consider Jason Stanley's attempt to subsume knowledge-how under an account of de se (...) knowledge. I argue that there are crucial differences between de se knowledge and knowledge-how. Thus, this paper advances both the discussion of intellectualism and the literature on the nature of de se knowledge. (shrink)
William Roche and Elliott Sober argue that explanatoriness is evidentially irrelevant. This conclusion is surprising since it conflicts with a plausible assumption—the fact that a hypothesis best explains a given set of data is evidence that the hypothesis is true. We argue that Roche and Sober's screening-off argument fails to account for a key aspect of evidential strength: the weight of a body of evidence. The weight of a body of evidence affects the resiliency of probabilities in the light of (...) new evidence. Thus, Roche and Sober are mistaken. Explanatoriness is evidentially relevant. (shrink)
We know facts, but we also know how to do things. To know a fact is to know that a proposition is true. But does knowing how to ride a bike amount to knowledge of propositions? This is a challenging question and one that deeply divides the contemporary landscape. A Critical Introduction to Knowledge-How introduces, outlines, and critically evaluates various contemporary debates surrounding the nature of knowledge-how. Carter and Poston show that situating the debate over the nature of knowledge-how in (...) other epistemological debates provides new ways to make progress. In particular, Carter and Poston explore the question of what knowledge-how involves, and how it might come apart from propositional knowledge, by engaging with key epistemological topics including epistemic luck, knowledge of language, epistemic value, virtue epistemology and social epistemology. New frontiers for research on knowledge-how are also explored relating to the internalism - externalism debate as well as embodied and extended knowledge. A Critical Introduction to Knowledge-How provides an accessible introduction to the main arguments in this important and thriving debate suited for undergraduates and postgraduates in philosophy and related areas. A strength of the book is its methodology which places a premium on placing the debates over knowledge-how in a broader conversation over the nature of knowledge. This book also offers an opinionated discussion of various lines of argument which will be of interest to professional philosophers as well. (shrink)
In this paper we argue that attention to the intricacies relating to belief illustrate crucial difficulties with Schellenberg's hiddenness argument. This issue has been only tangentially discussed in the literature to date. Yet we judge this aspect of Schellenberg's argument deeply significant. We claim that focus on the nature of belief manifests a central flaw in the hiddenness argument. Additionally, attention to doxastic subtleties provides important lessons about the nature of faith.
Twenty philosophers offer new essays examining the form of reasoning known as inference to the best explanation - widely used in science and in our everyday lives, yet still controversial. Best Explanations represents the state of the art when it comes to understanding, criticizing, and defending this form of reasoning.
It is a widespread intuition that the coherence of independent reports provides a powerful reason to believe that the reports are true. Formal results by Huemer, M. 1997. “Probability and Coherence Justification.” Southern Journal of Philosophy 35: 463–72, Olsson, E. 2002. “What is the Problem of Coherence and Truth?” Journal of Philosophy XCIX : 246–72, Olsson, E. 2005. Against Coherence: Truth, Probability, and Justification. Oxford University Press., Bovens, L., and S. Hartmann. 2003. Bayesian Epistemology. Oxford University Press, prove that, under (...) certain conditions, coherence cannot increase the probability of the target claim. These formal results, known as ‘the impossibility theorems’ have been widely discussed in the literature. They are taken to have significant epistemic upshot. In particular, they are taken to show that reports must first individually confirm the target claim before the coherence of multiple reports offers any positive confirmation. In this paper, I dispute this epistemic interpretation. The impossibility theorems are consistent with the idea that the coherence of independent reports provides a powerful reason to believe that the reports are true even if the reports do not individually confirm prior to coherence. Once we see that the formal discoveries do not have this implication, we can recover a model of coherence justification consistent with Bayesianism and these results. This paper, thus, seeks to turn the tide of the negative findings for coherence reasoning by defending coherence as a unique source of confirmation. (shrink)
There is an interesting and instructive problem with Richard Fumerton's acquaintance theory of noninferential justification. Fumerton's explicit account requires acquaintance with the truth-maker of one's belief and yet he admits that one can have noninferential justification when one is not acquainted with the truthmaker of one's belief but instead acquainted with a very similar truth-maker. On the face of it this problem calls for clarification. However, there are skeptical issues lurking in the background. This paper explores these issues by developing (...) a dilemma for an acquaintance theory. (shrink)
Although inference to the best explanation is ubiquitous in science and our everyday lives, there are numerous objections to the viability of IBE. Many of these objections have been thoroughly discussed, however, at least one objection to IBE has not received adequate treatment. We term this objection the “Disjunction Objection”. This objection challenges IBE on the grounds that it seems that even if H is the best explanation, it could be that the disjunction of its rivals is more likely to (...) be true. As a result, IBE appears to license accepting a hypothesis that is more likely than not to be false. Despite these initial appearances, we argue that the Disjunction Objection fails to impugn IBE. (shrink)
This paper responds to Ernest Sosa's recent criticism of Richard Fumerton's acquaintance theory. Sosa argues that Fumerton's account of non-inferential justification falls prey to the problem of the speckled hen. I argue that Sosa's criticisms are both illuminating and interesting but that Fumerton's theory can escape the problem of the speckled hen. More generally, the paper shows that an internalist account of non-inferential justification can survive the powerful objections of the Sellarsian dilemma and the problem of the speckled hen.
Bradford Hill (1965) highlighted nine aspects of the complex evidential situation a medical researcher faces when determining whether a causal relation exists between a disease and various conditions associated with it. These aspects are widely cited in the literature on epidemiological inference as justifying an inference to a causal claim, but the epistemological basis of the Hill aspects is not understood. We offer an explanatory coherentist interpretation, explicated by Thagard's ECHO model of explanatory coherence. The ECHO model captures the complexity (...) of epidemiological inference and provides a tractable model for inferring disease causation. We apply this model to three cases: the inference of a causal connection between the Zika virus and birth defects, the classic inference that smoking causes cancer, and John Snow’s inference about the cause of cholera. (shrink)
This paper develops and defends a coherentist account of reasons. I develop three core ideas for this defense: a distinction between basic reasons and noninferential justification, the plausibility of the neglected argument against first philosophy, and an emergent account of reasons. These three ideas form the backbone for a credible coherentist view of reasons. I work toward this account by formulating and explaining the basic reasons dilemma. This dilemma reveals a wavering attitude that coherentists have had toward basic reasons. More (...) importantly, the basic reasons dilemma focuses our attention on the central problems that afflict coherentist views of basic beliefs. By reflecting on the basic reasons dilemma, I formulate three desiderata that any viable coherentist account of basic beliefs must satisfy. I argue that the account on offer satisfies these desiderata. (shrink)
Functionalism about truth is the view that truth is an explanatorily significant but multiply-realizable property. According to this view the properties that realize truth vary from domain to domain, but the property of truth is a single, higher-order, domain insensitive property. We argue that this view faces a challenge similar to the one that Jaegwon Kim laid out for the multiple realization thesis. The challenge is that the higher-order property of truth is equivalent to an explanatorily idle disjunction of its (...) realization bases. This consequence undermines the alethic functionalists’ non-deflationary ambitions. A plausible response to Kim’s argument fails to carry over to alethic functionalism on account of significant differences between alethic functionalism and psychological functionalism. Lynch’s revised view in his book Truth as One and Many ( 2009 ) fails to answer our challenge. The upshot is that, while mental functionalism may survive Kim’s argument, it mortally wounds functionalism about truth. (shrink)
Ted Sider’s paper “Hell and Vagueness” challenges a certain conception of Hell by arguing that it is inconsistent with God’s justice. Sider’s inconsistencyargument works only when supplemented by additional premises. Key to Sider’s case is a premise that the properties upon which eternal destinies superveneare “a smear,” i.e., they are distributed continuously among individuals in the world. We question this premise and provide reasons to doubt it. The doubts come from two sources. The first is based on evidential considerations borrowed (...) from skeptical theism. A related but separate consideration is that supposing it would be an insurmountable problem for God to make just (and therefore non-arbitrary) distinctions in morally smeared world, God thereby has sufficient motivation not to actualize such worlds. Yet God also clearly has motivation only to actualize some member of the subset of non-smeared worlds which don’t appear non-smeared. For if it was obvious who was morally fit for Heaven and who wasn’t, a new arena of great injustice is opened up. The result is that if there is a God, then he has the motivation and the ability to actualize from just that set of worlds which are not smeared but which are indiscernible from smeared worlds. (shrink)
Social Evil.Ted Poston - 2014 - Oxford Studies in Philosophy of Religion 5:209-233.details
Social evil is any pain or suffering brought about by game-theoretic interactions of many individuals. This paper introduces and discusses the problem of social evil. I begin by focusing on social evil brought about by game-theoretic interactions of rational moral individuals. The problem social evil poses for theism is distinct from problems posed by natural and moral evils. Social evil is not a natural evil because it is brought about by the choices of individuals. But social evil is not a (...) form of moral evil because each individual actor does not misuse his free will. Traditional defenses for natural and moral evil fall short in addressing the problem of social evil. The final section of this paper discusses social evil and virtue. I begin by arguing that social evil can arise even when individual virtue is lacking. Next, I explore the possibility of an Edwardsian defense of social evil that stresses the high demands of true virtue. In this context, I argue that social evil may arise even when all the participants are truly virtuous. The conclusion of this paper is that social evil is problematic and provides new ground for exploring the conceptual resources of theism. (shrink)
Chalmers (The character of consciousness, 2010) argues for an acquaintance theory of the justification of direct phenomenal beliefs. A central part of this defense is the claim that direct phenomenal beliefs are cognitively significant. I argue against this. Direct phenomenal beliefs are justified within the specious present, and yet the resources available with the present ‘now’ are so impoverished that it barely constrains the content of a direct phenomenal belief. I argue that Chalmers’s account does not have the resources for (...) explaining how direct phenomenal beliefs support the inference from ‘this E is R’ to ‘that was R.’. (shrink)
The coherence of independent reports provides a strong reason to believe that the reports are true. This plausible claim has come under attack from recent work in Bayesian epistemology. This work shows that, under certain probabilistic conditions, coherence cannot increase the probability of the target claim. These theorems are taken to demonstrate that epistemic coherentism is untenable. To date no one has investigated how these results bear on different conceptions of coherence. I investigate this situation using Thagard’s ECHO model of (...) explanatory coherence. Thagard’s ECHO model provides a natural representation of the evidential significance of multiple independent reports. (shrink)
The Sellarsian dilemma is a powerful argument against internalistic foundationalist views that aim to end the regress of reasons in experiential states. LaurenceBonJour once defended the soundness of this dilemma as part of a larger argument for epistemic coherentism. BonJour has now renounced his earlier conclusions about the dilemma and has offered an account of internalistic foundationalism aimed, in part, at showing the errors of his former ways. I contend that BonJour’s early concerns about the Sellarsian dilemma are correct, and (...) that his latest position does not adequately handle the dilemma. I focus my attention on BonJour’s claim that a nonconceptual experiential state can provide a subject with a reason to believe some proposition. It is crucial for the viability of internalistic foundationalism to evaluate whether this claim is true. I argue it is false. The requirement that the states that provide justification give reasons to a subject conflicts with the idea that these states are nonconceptual. In the final section I consider David Chalmers’s attempt to defend a view closely similar to BonJour’s. Chalmers’s useful theory of phenomenal concepts provides a helpful framework for identifying a crucial problem with attempts to end the regress of reasons in pure experiential states. (shrink)
Epistemic conservatism is the thesis that the mere holding of a belief confers some positive epistemic status on its content. Conservatism is widely criticized on the grounds that it conflicts with the main goal in epistemology to believe truths and disbelieve falsehoods. In this paper I argue for conservatism and defend it from objections. First, I argue that the objection to conservatism from the truth goal in epistemology fails. Second, I develop and defend an argument for conservatism from the perspectival (...) character of the truth goal. Finally, I examine several forceful challenges to conservatism and argue that these challenges are unsuccessful. The first challenge is that conservatism implies the propriety of assertions like ‘I believe p and this is part of my justification for it’. The second challenge argues that conservatism wrongly implies that the identity of an epistemic agent is relevant to the main goal of believing truths and disbelieving falsehoods. The last two challenges I consider are the ‘extra boost’ objection and the conversion objection. Each of these objections helps to clarify the nature of the conservative thesis. The upshot of the paper is that conservatism is an important and viable epistemological thesis. (shrink)
Abstract This paper develops an explanationist treatment of the problem of the criterion. Explanationism is the view that all justified reasoning is justified in virtue of the explanatory virtues: simplicity, fruitfulness, testability, scope, and conservativeness. A crucial part of the explanationist framework is achieving wide reflective equilibrium. I argue that explanationism offers a plausible solution to the problem of the criterion. Furthermore, I argue that a key feature of explanationism is the plasticity of epistemic judgments and epistemic methods. The explanationist (...) does not offer any fixed judgments or methods to guide epistemic conduct; even the explanatory virtues themselves are subject to change. This feature of explanationism gives it an advantage over non-explanationist views that offer fixed epistemic judgments and epistemic methods. The final section of this paper responds to objections to explanationism. (shrink)
We argue that there is a tension between two types of design arguments-the fine-tuning argument (FTA) and the biological design argument (BDA). The tension arises because the strength of each argument is inversely proportional to the value of a certain currently unknown probability. Since the value of that probability is currently unknown, we investigate the properties of the FTA and BDA on different hypothetical values of this probability. If our central claim is correct this suggests three results: 1. It is (...) not very plausible that a cumulative case for theism include both the FTA and the BDA (with one possible qualification); 2. Self-organization scenarios do not threaten theism but in fact provide the materials for a good FTA. 3. A plausible design argument of one sort or another (either FTA or BDA) will be available for a wide variety of values of the key probability. (shrink)
This paper articulates a way to ground a relatively high prior probability for grand explanatory theories apart from an appeal to simplicity. I explore the possibility of enumerating the space of plausible grand theories of the universe by using the explanatory properties of possible views to limit the number of plausible theories. I motivate this alternative grounding by showing that Swinburne’s appeal to simplicity is problematic along several dimensions. I then argue that there are three plausible grand views—theism, atheism, and (...) axiarchism—which satisfy explanatory requirements for plausibility. Other possible views lack the explanatory virtue of these three theories. Consequently, this explanatory grounding provides a way of securing a nontrivial prior probability for theism, atheism, and axiarchism. An important upshot of my approach is that a modest amount of empirical evidence can bear significantly on the posterior probability of grand theories of the universe. (shrink)
This paper addresses the scatter problem for foundational evidentialism. Reflection on the scatter problem uncovers significant epistemological lessons. The scatter problem is evaluated in connection with Ernest Sosa’s use of the problem as an argument against foundational evidentialism. Sosa’s strategy is to consider a strong intuition in favor of internalism—the new evil demon problem, and then illustrate how a foundational evidentialist account of the new evil demon problem succumbs to the scatter problem. The goal in this paper is to evaluate (...) the force of the scatter problem. The main argument of the paper is that the scatter problem has mixed success. On the one hand, scatter undermines objectual evidentialism, an evidentialist theory that formulates principles of basic perceptual justification in terms of the objects of perceptual states. On the other hand, the problem of scatter does not undermine content evidentialism, an evidentialist view that formulates its epistemic principles in terms of the assertive content of perceptual states. The significance of the scatter problem, especially in concert with the new evil demon problem, is that it provides an argument for content evidentialism. (shrink)
Timothy Williamson’s anti luminosity argument has received considerable attention. Escaping unnoticed, though, is a strikingly similar argument from David Hume. This paper highlights some of the arresting parallels between Williamson’s reasoning and Hume’s that will allow us to appreciate more deeply the plausibility of Williamson’s reasoning and to understand how, following Hume, we can extend this reasoning to undermine the “luminosity” of simple necessary truths. More broadly the parallels help us to identify a common skeptical predicament underlying both arguments, which (...) we shall call “the quarantine problem”. The quarantine problem expresses a deep skepticism about achieving any exalted epistemic state. Further, the perspective gained by the quarantine problem allows us to easily categorize existing responses to Williamson’s anti luminosity argument and to observe the deficiencies of those responses. In sum, the quarantine problem reveals the deeply fallibilistic nature of whatever knowledge we may possess. (shrink)
Epistemic conservatism is the thesis that the mere holding of a belief confers some positive epistemic status on its content. Conservatism is widely criticized on the grounds that it conflicts with the main goal in epistemology to believe truths and disbelieve falsehoods. In this paper I argue for conservatism and defend it from objections. First, I argue that the objection to conservatism from the truth goal in epistemology fails. Second, I develop and defend an argument for conservatism from the perspectival (...) character of the truth goal. Finally, I examine several forceful challenges to conservatism and argue that these challenges are unsuccessful. The first challenge is that conservatism implies the propriety of assertions like ‘I believe p and this is part of my justification for it’. The second challenge argues that conservatism wrongly implies that the identity of an epistemic agent is relevant to the main goal of believing truths and disbelieving falsehoods. The last two challenges I consider are the ‘extra boost’ objection and the conversion objection. Each of these objections helps to clarify the nature of the conservative thesis. The upshot of the paper is that conservatism is an important and viable epistemological thesis. (shrink)
Laurence BonJour's (1985) coherence theory of empirical knowledge relies heavily on a traditional foundationalist theory of a priori knowledge. He argues that a foundationalist, rationalist theory of a priori justification is indispensable for a coherence theory. BonJour (1998) continues this theme, arguing that a traditional account of a priori justification is indispensable for the justification of putative a priori truths, the justification of any non-observational belief and the justification of reasoning itself. While BonJour's indispensability arguments have received some critical discussion (...) (Gendler 2001; Harman 2001; Beebe 2008), no one has investigated the indispensability arguments from a coherentist perspective. This perspective offers a fruitful take on BonJour's arguments, because he does not appreciate the depth of the coherentist alternative to the traditional empiricist-rationalist debate. This is surprising on account of BonJour's previous defense of coherentism. Two significant conclusions emerge: first, BonJour's indispensability arguments beg central questions about an explanationist form of coherentism; second, BonJour's original defense of coherentism took on board certain assumptions that inevitably led to the demise of his form of coherentism. The positive conclusion of this article is that explanatory coherentism is more coherent than BonJour's indispensability arguments assume, and more coherent than BonJour's earlier coherentist epistemology. (shrink)
This article develops a theory of reasons that has strong similarities to Peter Klein's infinitism. The view it develops, Framework Reasons, upholds Klein's principles of avoiding arbitrariness (PAA) and avoiding circularity (PAC) without requiring an infinite regress of reasons. A view of reasons that holds that the “reason for” relation is constrained by PAA and that PAC can avoid an infinite regress if the “reason for” relation is contextual. Moreover, such a view of reasons can maintain that skepticism is false (...) by the maintaining that there is more to epistemic justification than can be expressed in any reasoning session. One crucial argument for Framework Reasons is that justification depends on a background of plausibility considerations. The final section of the article applies this view of reasons to Michael Bergmann's argument that any nonskeptical epistemology must embrace epistemic circularity. (shrink)
The thirteen newly commissioned essays in _The Mystery of Skepticism: New Explorations_ represent the cutting-edge of research on underexplored skeptical challenges, dimensions of the skeptical problematic, and responses to various kinds of skepticism.
The thirteen newly commissioned essays in _The Mystery of Skepticism: New Explorations_ represent the cutting-edge of research on underexplored skeptical challenges, dimensions of the skeptical problematic, and responses to various kinds of skepticism.