Skip to main content
Log in

A normative comparison of threshold views through computer simulations

  • Original Research
  • Published:
Synthese Aims and scope Submit manuscript

Abstract

The threshold view says that a person forms an outright belief P if and only if her credence for P reaches a certain threshold. Using computer simulations, I compare different versions of the threshold view to understand how they perform under time pressure in decision problems. The results illuminate the strengths and weaknesses of the various cognitive strategies in different decision contexts. A threshold view that performs well across diverse contexts is likely to be a cognitively flexible and context-dependent fusion of several of the existing theories. The results of the simulations also cast doubts on the possibility of a threshold view that is both simple enough to streamline our reasoning while also allowing us to form good action-guiding beliefs.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Fig. 4
Fig. 5

Similar content being viewed by others

Data availability

N/A.

Code availability

Code for the model can be found here: https://github.com/alicecwhuang/threshold_code.git.

Notes

  1. The lottery paradox is a problem for most threshold views with \(t < 1\). Suppose there is a lottery with one winning ticket out of n tickets, where \((1-\frac{1}{n}) > t\). For each individual ticket, S’s credence that it loses suffices for an outright belief according to standard threshold views. Therefore, S outright believes that each ticket will lose. However, S also outright believes, by assumption, that at least one ticket will win. It follows that conjunctive closure is violated. In other words, if \(t < 1\), then S outright believes P and Q without believing \(P\wedge Q\).

  2. Justifications for this assumption can be found in our linguistic practices (Maher, 1993), as well as from the perspective of cognitive efficiency, which I explain below. Moreover, see Weisberg (2020) for arguments from the perspective of psyontology that we have both binary and graded beliefs.

  3. That outright beliefs are at least partially fixed by credences is the core idea of the threshold view. Alternatives to the threshold view include, first, theories according to which graded and binary beliefs are entirely independent. This leads to prima facie strange results, for in virtue of what do the two systems both count as belief states if there is no connection between them? Secondly, graded beliefs might be fixed by binary beliefs. This approach is favored by Holton (2014) and Easwaran (2016), who think that graded beliefs are reducible to binary beliefs.

  4. Similar principles establishing the link between knowledge and action are endorsed in Fantl and McGrath (2002), Hawthorne (2003); Hawthorne and Stanley (2008), and Ross and Schroeder (2014).

  5. Greco (2015) later turns to support what I will call “SmallSens(1), ” or “credences-one sensitivism” and accept irrationality as an integral part of beliefs.

  6. Being subjected to a diachronic Dutch-Book might not be a mark of irrationality in some views. See for example Christensen (1991).

  7. Weatherson (2005) has a similar spirit of anti-idealization, though he focuses on limiting the domain where conjunctive closure should be obeyed, as opposed to limiting the set of active possibilities. He argues that we can have probabilistic coherence without logical coherence, and in particular, that only the set of salient propositions must be conjunctively closed. See also Weisberg (2020) for more on how credence is assigned on the fly, and Stanley (2007) for a normative construal of active possibilities.

  8. See Mayseless and Kruglanski (1987) for the psychological perspective on this topic.

  9. Babic (2019) endorses a pragmatic view similar to small domain purism, though he does not address the belief-credence connection, nor does he explicitly mention the concept of an active domain.

  10. See Nagel (2010) for an extensive discussion about the termination of inquiries.

  11. See Duckworth et al. (2018) 2.2.1 for a brief introduction.

  12. The Weibull distribution is determined by two parameters—shape and scale. The shape parameter is randomized between 1 and 2, and the scale is fixed at 50. The two parameters together determine, roughly, how rapidly the patient’s probability of death increases with each time step. For instance, when the shape parameter is small, the probability of death increases faster in the beginning and slower towards the end. The smaller the scale is, the faster the patients’ probability of death increases with time. My setup of the problem here roughly follows that of Douven (2020).

  13. The effects of correct and incorrect interventions are determined by parameters a and b. Let the patient’s baseline probability of death without intervention at a given time be p. The correct intervention brings down the probability of death to p/a, whereas a wrong intervention increases it to \(\frac{(p+1)}{b}\). The larger a is, the more effectively a correct intervention lowers the patient’s risk of death. The smaller b is, the more a wrong intervention increases the patient’s probability of death. a is randomized between 1.5 and 2.5 whereas b is fixed at 2.

  14. The symmetric dirichlet distribution is determined by a concentration parameter. The larger the concentration parameter is, the more equally probability is distributed across the hypotheses. By contrast, when the concentration parameter is small, the agent has high initial credences for a small number of possibilities and low credences for the other hypotheses. In the simulation, the concentration parameter is randomly chosen from a normal distribution centered at 0.5, with a standard deviation of 0.1. The reason for restricting the concentration parameter to a limited range is to prevent having priors too close to the uniform prior or too concentrated and thereby avoid giving small domain agents unfair advantage or disadvantage.

  15. This simulation can thus also be interpreted simply as a guessing game of coin biases with monetary rewards involved.

  16. Here I provide a more detailed example of the scoring rule. Suppose again that the doctor has a cluster infection of 3 patients, and the probability of death for each patient in the case at time t is 0.6. If the agent’s diagnosis is correct, she gains \((1-\frac{0.6}{a})\times 3\) points. If the intervention is incorrect, the agent gains \((1-\frac{1.6}{b})\times 3\) points, with a and b being the parameters explained in footnote 13. If after 100 time steps, no intervention has been made, then the agent gets \((1-p_{100})\times 3\) points, where \(p_{100}\) is the patient’s probability of death at the 100th time step.

  17. The normal distribution is truncated at 0 and 1.

  18. In the simulations, since the stakes range from 1 to 5, \(f(s)=\frac{s-3}{2}\). This yields the desired transformation with \(f(1) = -1\) and \(f(5) = 1\).

  19. When \(s=3\), the threshold is the default \(\tau \) regardless of sensitivity, since \(f(3) = 0\), as specified in footnote 18.

  20. Adjustments are made for cases where the domain size may be larger than N\(_{hyp}\).

  21. The code for the simulations can be found in Appendix C.

  22. Situations in which SmallSens(1) agents do form outright beliefs and take actions include (a) when the true coin bias is either 0 or 1, (b) when only one hypothesis in the active domain is neither 0 nor 1 (c) their credences are automatically rounded up to 1 by the computer, which happens at around the \(18^{th}\) decimal place.

  23. The bars are generated from data averaged over runs with N\(_{hyp} = 5, 6\) for easy problems and over runs with N\(_{hyp} = 14, 15\) for difficult problems.

  24. A related observation not shown in this graph is that the performances of small domain strategies are less affected by difficulty than full domain strategies. However, since the more possible alternatives there are, the more likely the truth is outside of the most probable \(4\pm 2\) hypotheses, there is still a negative correlation between the average scores and difficulty for small domain agents.

  25. The results do not render FullSens, specifically in high stake situations, a failed strategy tout court, since there might be situations where errors are so severely penalized that FullSens agents’ level of prudence in high-stake scenarios is warranted. This is not the case in the present setup.

  26. This suggestion requires the further assumption that an agent knows the costs of type 1 and type 2 errors, as well as how the costs change with time.

References

Download references

Acknowledgements

I am grateful to Jonathan Weisberg, Jennifer Nagel, Boris Babic as well as two anonymous referees for their helpful comments on various drafts of this paper. Special thanks are also due to Titas Geryba for much-appreciated copy-editing and insightful suggestions. Thanks also to Nischal Mainali for interesting interdisciplinary discussions about cognitive constraints.

Funding

None.

Author information

Authors and Affiliations

Authors

Contributions

N/A.

Corresponding author

Correspondence to Alice C. W. Huang.

Ethics declarations

Conflict of interest

The authors declare that they have no conflict of interest.

Additional information

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Appendices

Appendix A: Weibull distribution

The Weibull distribution is a continuous probability distribution defined by the function

$$\begin{aligned} f(x; \lambda , k)={\left\{ \begin{array}{ll} \frac{k}{\lambda }(\frac{x}{\lambda })^{k-1}e^{-(x/\lambda )^k},&{}\quad \text {if }x \le 0.\\ 0,&{}\quad \text {otherwise}, \end{array}\right. } \end{aligned}$$
(2)

where \(\lambda \in (0, +\infty )\) is the scale parameter and \(k\in (0, +\infty )\) is the shape parameter. The cumulative distribution function of the Weibull distribution is given by

$$\begin{aligned} f(x; \lambda , k)={\left\{ \begin{array}{ll} 1-e^{-(x/\lambda )^k},&{}\quad \text {if }x \le 0.\\ 0,&{}\quad \text {otherwise}. \end{array}\right. } \end{aligned}$$
(3)

Appendix B: Symmetric dirichlet distribution

The Dirichlet distribution is a multivariate generalization of the beta distribution, parameterized by a vector \(\varvec{\alpha }\) with real entries. The Dirichlet distribution with parameters \(\alpha _1, \alpha _2, \ldots , \alpha _K\) is given by

$$\begin{aligned} f(x_1, x_2, \ldots , x_K; \alpha _1, \alpha _2, \ldots , \alpha _K) = \frac{1}{B(\varvec{\alpha })}\prod _{i = 1}^{K} x_{i}^{\alpha _{i}-1}, \end{aligned}$$
(4)

where \(\sum _{i = 1}^{K} x_i = 1\) and \(x_1, x_2, \ldots , x_K > 0\). The normalizing constant \(B(\varvec{\alpha })\) is the multivariate beta function

$$\begin{aligned} B(\varvec{\alpha }) = \frac{\prod _{i = 1}^{K} \Gamma (\alpha _i)}{\Gamma \left( \sum _{i = 1}^{K}\alpha _i\right) }, \end{aligned}$$
(5)

where \(\Gamma \) is the gamma function.

A special case of the Dirichlet distribution is the symmetric Dirichlet Distribution. In a symmetric Dirichlet distribution, the entries of the vector \(\varvec{\alpha }\) are all equal. This corresponds to the case where there is no prior information that leads us to prefer one possibility over another. Since all the entries of \(\varvec{\alpha }\) are the same, the symmetric Dirichlet distribution can simply be parameterized by a scalar value \(\alpha \), often called the concentration parameter. When \(\alpha > 1\) the probability is randomly but evenly distributed across the possibilities. When \(\alpha < 1\), most of the probability will be concentrated in a few of the possibilities. The symmetric Dirichlet distribution is given by

$$\begin{aligned} f(x_1, x_2, \ldots , x_K; \alpha ) = \frac{\Gamma (\alpha K)}{\Gamma (\alpha )^K}\prod _{i = 1}^{K} x_{i}^{\alpha _{i}-1}. \end{aligned}$$
(6)

Appendix C: Code for the model

The code for the model, written in Python using the agent-based simulation package Mesa, can be found here: https://github.com/alicecwhuang/threshold_code.git.

Rights and permissions

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Huang, A.C.W. A normative comparison of threshold views through computer simulations. Synthese 200, 296 (2022). https://doi.org/10.1007/s11229-022-03784-x

Download citation

  • Received:

  • Accepted:

  • Published:

  • DOI: https://doi.org/10.1007/s11229-022-03784-x

Keywords

Navigation