Skip to main content
Log in

Counterfeit testimony: lies, trust, and the exchange of information

  • Published:
Philosophical Studies Aims and scope Submit manuscript

Abstract

Most explanations of the rational authority of testimony provide little guidance when evaluating individual pieces of testimony. In practice, however, we are remarkably sensitive to the varying epistemic credentials of testimony: extending trust when it is deserved, and withholding it when it is not. A complete account of the epistemology of testimony should, then, have something to say about when it is that testimony is trustworthy. In the typical case, to judge someone trustworthy requires judging them to be competent and sincere. In this essay we develop an exchange-based account of testimony that shows how those who receive testimony are in a position to evaluate the sincerity of speakers.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Similar content being viewed by others

Notes

  1. This is the tack taken by Fricker (1987). For the outlines of a different kind of reductive view, see Dogramaci (2012), and for commentary on it see (Tebben and Waterman 2015a).

  2. Although there are, of course, variations within both camps. For good examples of non-reductionist views see Graham (2006) and Coady (1992). Faulkner (2011) says that his view is a hybrid position, although he agrees with the non-reductionists on their central contention.

  3. Lackey (2006: 173). Faulkner has also urged that considerations of these sorts provide some means of distinguishing reliable from unreliable testimony. See, for example, Faulkner (2002).

  4. We would also note that it might not be quite fair to criticize standard reductionist (or anti-reductionist) views for failing to provide guidance on which instances of testimony to accept. They may simply have other (more theoretical and less practical) aims. Nevertheless, even if extant theories do not aim to provide such guidance, it is certainly an advance if this guidance can be offered.

  5. Dogramaci (2012) argues that, since epistemic agents trust their own belief-forming procedures, one way to see others as competent is by seeing others as sharing those belief-forming procedures. We think that this is on the right track. In our earlier work we argued that epistemic agents’ need to keep a store of information with high exchange value encourages everyone in a community to use the same belief-forming procedures. If so, for anyone that an agent is likely to encounter, there is reason to believe that that person reasons as he or she does, hence in a way that he or she takes to be reliable. [See Tebben and Waterman (2015a, b), as well as Graham et al. (2015)].

  6. Some successful artists have even given away their recordings for free, essentially letting them serve as advertising for their live performances. Two examples: Nine Inch Nails charged nothing for downloads of their album The Slip, and Radiohead permitted listeners to download copies of In Rainbows for any amount that they would like to pay, including nothing.

  7. Some important caveats apply. We can, in small groups, monitor the distribution of information. If Carla tells Fraser a secret, and enjoins him not to share it, then Fraser has an incentive to stay quiet for fear of losing Carla’s trust. The spread of information is also linked to the degree to which the information is indirectly rivalrous. If the information in question concerns the location of an apple (which is otherwise available to be eaten), one should expect audience members to be more reluctant to spread this information than if the information at issue is not indirectly rivalrous.

  8. And, indeed, by some non-traditional reductionistic theories as well. For example, Dogramaci’s “epistemic communism” (see Dogramaci 2012).

  9. It is, for example, iteration of the game that allows the famous tit-for-tat strategy to do well across a range of competing strategies in a prisoner’s dilemma game (see Axelrod 1984).

  10. As noted in the main text, this is one of Aesop’s fables. We do not know if there is a canonical version of it; the version presented here is simply the story, as best as we can recall it, as it was told to us when we were children.

  11. Say that, in fact, if the boy falsely cries ‘WOLF!’ x-many times, the villagers will stop responding to his signals. In order to maximize his payoff, the best strategy for the boy is actually to falsely cry ‘WOLF!’ x − 1 times. Fortunately for parents who are trying to teach their children not to lie, children rarely pick up on this.

  12. Part of the reason that they are rare is that, when punishment is costly, they give rise to second-order free rider problems. If either A or B could punish C for lying, it is rational for A to rely on B to pay the cost of punishing C, and for B to rely on A to do it. On this topic see Boyd and Richerson (1992: 174).

  13. We focus on what we call the “ordinary case” of exchange, where interlocutors are both part of a single epistemic community and hence can influence each-others’ reputations. The ordinary case is what game theorists would call an iterated game. Though much less common than the ordinary case, testimony is sometimes exchanged by individuals who are not part of the same epistemic community, as when we ask for directions in an unfamiliar city. Cases like these, which game theorists call one-off games, introduce some special considerations that are discussed in Sect. 4.

  14. Vigilance involves more than trust: we are sensitive to an agent’s competence as well as their intentions and reputation for sincerity (Sperber et al. 2010: 383). There is a lively empirical debate over the extent of our vigilance (see Gilbert 1993), but we believe that the weight of the empirical evidence is on the side of Sperber, et al. To their formula we would add that agents are also sensitive to the utility of the information being exchanged.

  15. To a first approximation the discount will be a function of how often we expect a speaker to lie, and the average value of the information they have on offer. It might seem like we should not accept information from a known liar, but this is not the case so long as it’s possible to estimate how often they lie. Take for example the car dealer who is known to sell lemons: if one in five cars priced at $20 k is a worthless lemon, then it is rational to buy from the dealer at the discounted price of $16 k.

  16. Do lies lose value when they’re passed on, in the way that truthful testimony does? In some special cases perhaps. In those cases a term denoting this lost value should be added to the left side of (2). However, we believe these will be marginal cases. Typically the utility of a lie is derived from affecting the behavior of a particular individual. When we (falsely) tell the credit card company that the bill is in the mail, few people are positioned to benefit from this lie. Often no one else cares, and the speaker has little to gain or lose from the spread of the information.

  17. It has been suggested to us that this argument only works if Norm does not have good reason to believe that Cliff is honest. If he has good reason to think him honest, the suggestion goes, he will disregard what Lilith tells him. We’re not so sure. If he has good (though in this case misleading) reason to believe that Lilith is reliable, he may disregard the evidence that he possesses. If he takes her to be about as reliable as he is himself, the effect of Lilith’s testimony depends on how problems of peer disagreement are resolved. (A contentious issue that we will not address here).

  18. No one need consciously go in for this reasoning. It could be individually conditioned. Or (perhaps more plausibly) societies in which information is inexpensive tend to adapt better to changing circumstances, and so the, culturally instituted, practice of offering reliable gossip could spread because it results in more adaptable societies. We won’t speculate on this matter any further.

  19. It is worth remarking that this information is usually not very valuable. It tends to be either widely available or not indirectly rivalrous. Speakers happily give away directions to the grocery store, but they tend to be more reticent about their bank PIN numbers.

  20. This problem is well known to economists and game theorists. See Dasgupta (1988: 66 n 20), and the sources cited therein.

  21. For a nice presentation of this solution, see McElreath and Boyd (2007, chapter 4). They credit Axelrod and Hamilton with the idea.

  22. There is, interestingly, considerable empirical evidence that we make extremely rapid, intuitive judgments about the trustworthiness of unfamiliar individuals based on their appearance, particularly facial expressions (Willis and Todorov 2006; Ybarra et al. 2001).

  23. Compare this to act utilitarians’ acceptance of “rules of thumb” for making decisions when a lack of time or information precludes doing a full utilitarian calculation.

  24. It is worth remarking that the risk of deception for the sake of manipulation (see Sperber et al. 2010; Mercier and Sperber 2011) is low in one-off games. Participants in one-off games rarely have enough information about each other to permit manipulation through dishonesty.

References

  • Axelrod, R. (1984). The evolution of cooperation. Basic Books. (No city of publication given.)

  • Boyd, R., & Richerson, P. J. (1992). Punishment allows the evolution of cooperation (or anything else) in sizable groups. Ethology and Sociobiology, 13, 171–195.

    Article  Google Scholar 

  • Boyd, R., & Richerson, P. J. (2005). The origins and evolution of cultures. New York: Oxford University Press.

    Google Scholar 

  • Clément, F., Koenig, M., & Harris, P. (2004). The ontogeny of trust. Mind and Language, 19, 360–379.

    Article  Google Scholar 

  • Coady, C. A. J. (1992). Testimony: A philosophical study. Oxford: Oxford University Press.

    Google Scholar 

  • Dasgupta, P. (1988). Trust as a commodity. In E. Gambetta (Ed.), Trust: Making and breaking cooperative relations (pp. 49–72). Oxford: Blackwell.

    Google Scholar 

  • Dogramaci, S. (2012). Reverse engineering epistemic evaluations. Philosophy and Phenomenological Research, 84, 513–530.

    Article  Google Scholar 

  • Dunbar, R. (2004). Gossip in evolutionary perspective. Review of General Psychology, 9, 100–110.

    Article  Google Scholar 

  • Faulkner, P. (2002). On the rationality of our response to testimony. Synthese, 131, 353–370.

    Article  Google Scholar 

  • Faulkner, P. (2011). Knowledge on trust. Oxford: Oxford University Press.

    Book  Google Scholar 

  • Fricker, E. (1987). The epistemology of testimony. In Proceedings of the Aristotelian Society, supplementary volumes, 61, pp. 57–83 + 85–106.

  • Gilbert, D. T., Tafarodi, R. W., & Malone, P. S. (1993). You can’t not believe everything you read. Journal of Personality and Social Psychology, 65, 221–233.

    Article  Google Scholar 

  • Graham, P. (2006). Liberal fundamentalism and its rivals. In J. Lackey & E. Sosa (Eds.), The epistemology of testimony (pp. 93–115). Oxford: Oxford University Press.

    Chapter  Google Scholar 

  • Graham, P., Bachman, Z., McFadden, M., & Stotts, M. (2015). Epistemic evaluations: Consequences, costs and benefits. Social Epistemology Review and Reply Collective, 4(4), 7–13.

    Google Scholar 

  • Koenig, M., & Echols, C. (2003). Infants understanding of false labeling events: the referential roles of words and the speakers who use them. Cognition, 87, 179–203.

    Article  Google Scholar 

  • Lackey, J. (2006). It takes two to tango: beyond reductionism and non-reductionism. In J. Lackey & E. Sosa (Eds.), The epistemology of testimony (pp. 160–192). Oxford: Oxford University Press.

    Chapter  Google Scholar 

  • McElreath, R., & Boyd, R. (2007). Mathematical models of social evolution. Chicago: University of Chicago Press.

    Book  Google Scholar 

  • Mercier, H., & Sperber, D. (2011). Why do humans reason? arguments for an argumentative theory. Behavioral and Brain Sciences, 34, 57–111.

    Article  Google Scholar 

  • Nowak, M. (2006). Five rules for the evolution of cooperation. Science, 314, 1560–1563.

    Article  Google Scholar 

  • Pradel, J., Euler, H., & Fechenhauer, D. (2009). Spotting altruistic dictator game players and mingling with them: the elective assortation of classmates. Evolution and Human Behavior, 30, 103–113.

    Article  Google Scholar 

  • Sperber, D., CleMent, F., Heintz, C., Mascaro, O., Mercier, H., Origgi, G., & Wilson, D. (2010). Epistemic vigilance. Mind and Language, 25, 359–393.

    Article  Google Scholar 

  • Tebben, N., & Waterman, J. P. (2015a). Epistemic free riders and reasons to trust testimony. Social Epistemology, 29, 270–279.

    Article  Google Scholar 

  • Tebben, N., & Waterman, J. P. (2015b). The market for testimony: A reply to our critics. Social Epistemology Review and Reply Collective, 4(5), 43–51.

    Google Scholar 

  • Willis, J., & Todorov, A. (2006). First impressions: making up your mind after a 100-ms exposure to a face. Psychological Science, 17, 592–598.

    Article  Google Scholar 

  • Ybarra, O., Chan, E., & Park, D. (2001). Young and old adults’ concerns about morality and competence. Motivation and Emotion, 25, 85–100.

    Article  Google Scholar 

Download references

Acknowledgments

An earlier version of this paper was presented in Prague to a conference on the “Nature of Rules and Normativity”, hosted by the Academy of Sciences of the Czech Republic. We would like to thank the members of our audience at that event for their helpful feedback. We would also like to thank an anonymous reviewer for providing written comments.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Nicholas Tebben.

Rights and permissions

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Tebben, N., Waterman, J. Counterfeit testimony: lies, trust, and the exchange of information. Philos Stud 173, 3101–3117 (2016). https://doi.org/10.1007/s11098-016-0652-0

Download citation

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s11098-016-0652-0

Keywords

Navigation