Abstract
Most explanations of the rational authority of testimony provide little guidance when evaluating individual pieces of testimony. In practice, however, we are remarkably sensitive to the varying epistemic credentials of testimony: extending trust when it is deserved, and withholding it when it is not. A complete account of the epistemology of testimony should, then, have something to say about when it is that testimony is trustworthy. In the typical case, to judge someone trustworthy requires judging them to be competent and sincere. In this essay we develop an exchange-based account of testimony that shows how those who receive testimony are in a position to evaluate the sincerity of speakers.
Similar content being viewed by others
Notes
We would also note that it might not be quite fair to criticize standard reductionist (or anti-reductionist) views for failing to provide guidance on which instances of testimony to accept. They may simply have other (more theoretical and less practical) aims. Nevertheless, even if extant theories do not aim to provide such guidance, it is certainly an advance if this guidance can be offered.
Dogramaci (2012) argues that, since epistemic agents trust their own belief-forming procedures, one way to see others as competent is by seeing others as sharing those belief-forming procedures. We think that this is on the right track. In our earlier work we argued that epistemic agents’ need to keep a store of information with high exchange value encourages everyone in a community to use the same belief-forming procedures. If so, for anyone that an agent is likely to encounter, there is reason to believe that that person reasons as he or she does, hence in a way that he or she takes to be reliable. [See Tebben and Waterman (2015a, b), as well as Graham et al. (2015)].
Some successful artists have even given away their recordings for free, essentially letting them serve as advertising for their live performances. Two examples: Nine Inch Nails charged nothing for downloads of their album The Slip, and Radiohead permitted listeners to download copies of In Rainbows for any amount that they would like to pay, including nothing.
Some important caveats apply. We can, in small groups, monitor the distribution of information. If Carla tells Fraser a secret, and enjoins him not to share it, then Fraser has an incentive to stay quiet for fear of losing Carla’s trust. The spread of information is also linked to the degree to which the information is indirectly rivalrous. If the information in question concerns the location of an apple (which is otherwise available to be eaten), one should expect audience members to be more reluctant to spread this information than if the information at issue is not indirectly rivalrous.
And, indeed, by some non-traditional reductionistic theories as well. For example, Dogramaci’s “epistemic communism” (see Dogramaci 2012).
It is, for example, iteration of the game that allows the famous tit-for-tat strategy to do well across a range of competing strategies in a prisoner’s dilemma game (see Axelrod 1984).
As noted in the main text, this is one of Aesop’s fables. We do not know if there is a canonical version of it; the version presented here is simply the story, as best as we can recall it, as it was told to us when we were children.
Say that, in fact, if the boy falsely cries ‘WOLF!’ x-many times, the villagers will stop responding to his signals. In order to maximize his payoff, the best strategy for the boy is actually to falsely cry ‘WOLF!’ x − 1 times. Fortunately for parents who are trying to teach their children not to lie, children rarely pick up on this.
Part of the reason that they are rare is that, when punishment is costly, they give rise to second-order free rider problems. If either A or B could punish C for lying, it is rational for A to rely on B to pay the cost of punishing C, and for B to rely on A to do it. On this topic see Boyd and Richerson (1992: 174).
We focus on what we call the “ordinary case” of exchange, where interlocutors are both part of a single epistemic community and hence can influence each-others’ reputations. The ordinary case is what game theorists would call an iterated game. Though much less common than the ordinary case, testimony is sometimes exchanged by individuals who are not part of the same epistemic community, as when we ask for directions in an unfamiliar city. Cases like these, which game theorists call one-off games, introduce some special considerations that are discussed in Sect. 4.
Vigilance involves more than trust: we are sensitive to an agent’s competence as well as their intentions and reputation for sincerity (Sperber et al. 2010: 383). There is a lively empirical debate over the extent of our vigilance (see Gilbert 1993), but we believe that the weight of the empirical evidence is on the side of Sperber, et al. To their formula we would add that agents are also sensitive to the utility of the information being exchanged.
To a first approximation the discount will be a function of how often we expect a speaker to lie, and the average value of the information they have on offer. It might seem like we should not accept information from a known liar, but this is not the case so long as it’s possible to estimate how often they lie. Take for example the car dealer who is known to sell lemons: if one in five cars priced at $20 k is a worthless lemon, then it is rational to buy from the dealer at the discounted price of $16 k.
Do lies lose value when they’re passed on, in the way that truthful testimony does? In some special cases perhaps. In those cases a term denoting this lost value should be added to the left side of (2). However, we believe these will be marginal cases. Typically the utility of a lie is derived from affecting the behavior of a particular individual. When we (falsely) tell the credit card company that the bill is in the mail, few people are positioned to benefit from this lie. Often no one else cares, and the speaker has little to gain or lose from the spread of the information.
It has been suggested to us that this argument only works if Norm does not have good reason to believe that Cliff is honest. If he has good reason to think him honest, the suggestion goes, he will disregard what Lilith tells him. We’re not so sure. If he has good (though in this case misleading) reason to believe that Lilith is reliable, he may disregard the evidence that he possesses. If he takes her to be about as reliable as he is himself, the effect of Lilith’s testimony depends on how problems of peer disagreement are resolved. (A contentious issue that we will not address here).
No one need consciously go in for this reasoning. It could be individually conditioned. Or (perhaps more plausibly) societies in which information is inexpensive tend to adapt better to changing circumstances, and so the, culturally instituted, practice of offering reliable gossip could spread because it results in more adaptable societies. We won’t speculate on this matter any further.
It is worth remarking that this information is usually not very valuable. It tends to be either widely available or not indirectly rivalrous. Speakers happily give away directions to the grocery store, but they tend to be more reticent about their bank PIN numbers.
This problem is well known to economists and game theorists. See Dasgupta (1988: 66 n 20), and the sources cited therein.
For a nice presentation of this solution, see McElreath and Boyd (2007, chapter 4). They credit Axelrod and Hamilton with the idea.
Compare this to act utilitarians’ acceptance of “rules of thumb” for making decisions when a lack of time or information precludes doing a full utilitarian calculation.
References
Axelrod, R. (1984). The evolution of cooperation. Basic Books. (No city of publication given.)
Boyd, R., & Richerson, P. J. (1992). Punishment allows the evolution of cooperation (or anything else) in sizable groups. Ethology and Sociobiology, 13, 171–195.
Boyd, R., & Richerson, P. J. (2005). The origins and evolution of cultures. New York: Oxford University Press.
Clément, F., Koenig, M., & Harris, P. (2004). The ontogeny of trust. Mind and Language, 19, 360–379.
Coady, C. A. J. (1992). Testimony: A philosophical study. Oxford: Oxford University Press.
Dasgupta, P. (1988). Trust as a commodity. In E. Gambetta (Ed.), Trust: Making and breaking cooperative relations (pp. 49–72). Oxford: Blackwell.
Dogramaci, S. (2012). Reverse engineering epistemic evaluations. Philosophy and Phenomenological Research, 84, 513–530.
Dunbar, R. (2004). Gossip in evolutionary perspective. Review of General Psychology, 9, 100–110.
Faulkner, P. (2002). On the rationality of our response to testimony. Synthese, 131, 353–370.
Faulkner, P. (2011). Knowledge on trust. Oxford: Oxford University Press.
Fricker, E. (1987). The epistemology of testimony. In Proceedings of the Aristotelian Society, supplementary volumes, 61, pp. 57–83 + 85–106.
Gilbert, D. T., Tafarodi, R. W., & Malone, P. S. (1993). You can’t not believe everything you read. Journal of Personality and Social Psychology, 65, 221–233.
Graham, P. (2006). Liberal fundamentalism and its rivals. In J. Lackey & E. Sosa (Eds.), The epistemology of testimony (pp. 93–115). Oxford: Oxford University Press.
Graham, P., Bachman, Z., McFadden, M., & Stotts, M. (2015). Epistemic evaluations: Consequences, costs and benefits. Social Epistemology Review and Reply Collective, 4(4), 7–13.
Koenig, M., & Echols, C. (2003). Infants understanding of false labeling events: the referential roles of words and the speakers who use them. Cognition, 87, 179–203.
Lackey, J. (2006). It takes two to tango: beyond reductionism and non-reductionism. In J. Lackey & E. Sosa (Eds.), The epistemology of testimony (pp. 160–192). Oxford: Oxford University Press.
McElreath, R., & Boyd, R. (2007). Mathematical models of social evolution. Chicago: University of Chicago Press.
Mercier, H., & Sperber, D. (2011). Why do humans reason? arguments for an argumentative theory. Behavioral and Brain Sciences, 34, 57–111.
Nowak, M. (2006). Five rules for the evolution of cooperation. Science, 314, 1560–1563.
Pradel, J., Euler, H., & Fechenhauer, D. (2009). Spotting altruistic dictator game players and mingling with them: the elective assortation of classmates. Evolution and Human Behavior, 30, 103–113.
Sperber, D., CleMent, F., Heintz, C., Mascaro, O., Mercier, H., Origgi, G., & Wilson, D. (2010). Epistemic vigilance. Mind and Language, 25, 359–393.
Tebben, N., & Waterman, J. P. (2015a). Epistemic free riders and reasons to trust testimony. Social Epistemology, 29, 270–279.
Tebben, N., & Waterman, J. P. (2015b). The market for testimony: A reply to our critics. Social Epistemology Review and Reply Collective, 4(5), 43–51.
Willis, J., & Todorov, A. (2006). First impressions: making up your mind after a 100-ms exposure to a face. Psychological Science, 17, 592–598.
Ybarra, O., Chan, E., & Park, D. (2001). Young and old adults’ concerns about morality and competence. Motivation and Emotion, 25, 85–100.
Acknowledgments
An earlier version of this paper was presented in Prague to a conference on the “Nature of Rules and Normativity”, hosted by the Academy of Sciences of the Czech Republic. We would like to thank the members of our audience at that event for their helpful feedback. We would also like to thank an anonymous reviewer for providing written comments.
Author information
Authors and Affiliations
Corresponding author
Rights and permissions
About this article
Cite this article
Tebben, N., Waterman, J. Counterfeit testimony: lies, trust, and the exchange of information. Philos Stud 173, 3101–3117 (2016). https://doi.org/10.1007/s11098-016-0652-0
Published:
Issue Date:
DOI: https://doi.org/10.1007/s11098-016-0652-0