Skip to main content
Log in

Modelling mechanisms with causal cycles

  • Published:
Synthese Aims and scope Submit manuscript

Abstract

Mechanistic philosophy of science views a large part of scientific activity as engaged in modelling mechanisms. While science textbooks tend to offer qualitative models of mechanisms, there is increasing demand for models from which one can draw quantitative predictions and explanations. Casini et al. (Theoria 26(1):5–33, 2011) put forward the Recursive Bayesian Networks (RBN) formalism as well suited to this end. The RBN formalism is an extension of the standard Bayesian net formalism, an extension that allows for modelling the hierarchical nature of mechanisms. Like the standard Bayesian net formalism, it models causal relationships using directed acyclic graphs. Given this appeal to acyclicity, causal cycles pose a prima facie problem for the RBN approach. This paper argues that the problem is a significant one given the ubiquity of causal cycles in mechanisms, but that the problem can be solved by combining two sorts of solution strategy in a judicious way.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Institutional subscriptions

Fig. 1
Fig. 2
Fig. 3
Fig. 4
Fig. 5
Fig. 6
Fig. 7
Fig. 8
Fig. 9
Fig. 10
Fig. 11

Similar content being viewed by others

Notes

  1. Note that, as with all models, a RBN model only models some aspects of a mechanism. The main goal is to model the hierarchical structure of the mechanism together with the causal structure at each level of the hierarchy, in such a way that the model can be used to draw quantitative inferences. See Casini et al. (2011) for a fuller presentation of the motivation behind this sort of model, and Sect. 7 of this paper for pointers to possible limitations of the RBN approach.

  2. Woodward’s concept of ‘mechanism’, or more precisely: of ‘mechanistic model’, is not explicitly multi-level or hierarchical, in contrast to those on which we focus in this paper. In the next section, the hierarchical nature will serve as one of the main reasons to adopt the RBN approach to mechanistic modelling.

  3. As such, his account of causality can also form the starting point for causal Bayes net accounts of the structure of scientific theories (see Leuridan 2014).

  4. For mentions of causal feedback in Explaining the Brain, see e.g. pp. 81 and 178–180. Several of Craver’s figures also contain cycles: see Fig. 4.1 (p. 115) and Fig. 4.6 (p. 121) and relatedly 5.7 (p. 189) and 5.8 (p. 194). Moreover, Fig. 3.2 (p. 71) and relatedly 4.1 (p. 166), leave open the possibility of causal feedback.

  5. This property of negative feedback systems to tend toward equilibrium is the case when there is no significant delay in the system. When delay is present, as it is in many biological systems, oscillations will tend to arise. We would like to thank Mike Joffe for pressing us on this issue.

  6. However, this is not always the case. For instance, various positive feedback loops in pregnancy serve to appropriately maintain hormone levels.

  7. These kinds of contingent-feedback cycles are, in other words, more sensitive to background conditions than the other two kinds. This makes them unstable, in Mitchell’s sense of stability as describing the sensitivity of relations to their background conditions (Mitchell 2009, p. 56). This should be discriminated from robustness, which describes the degree to which a function is maintained when one or more constitutive elements are disrupted (Mitchell 2009, pp. 69–73).

  8. This cycle is more complicated than suggested above. For example, the thyroid secretes two hormones—T3 and T4—which can be interconverted, and feedback occurs at various intermediate points in the cycle. But this simple version is adequate for our discussion.

  9. Similar examples of neurological oscillators are also discussed by (Bechtel (2011), pp. 548–549).

  10. Structural equation modelling had previously been put forward for this purpose. But structural equation models attempt to model deterministic relationships between cause and effect (with error terms which are usually assumed to be independent and normally distributed), while Bayesian networks seek to represent the probability distribution of the variables in question. In general it is harder to devise an accurate model of deterministic relationships than it is to determine probabilistic relationships between cause and effect.

  11. Note that RBNs are not the only hierarchical extension of Bayesian nets. See (Williamson (2005), §10.2) for a comparison between RBNs and other related formalisms.

  12. Although Bayesian nets have been extended to handle certain continuous cases, we restrict attention to discrete variables in this paper.

  13. This corresponds to the notion of ‘bottoming-out’ in the mechanistic literature (see Sect. 2).

  14. The malfunctioning of mechanisms is of particular interest to, e.g., neuroscience (Craver 2007, pp. 124–125) and medicine (Nervi 2010).

  15. While these arrows would not normally be interpreted causally, the question arises as to whether they might be if Craver’s views of mutual manipulability and causality are endorsed. See footnote 25.

  16. In this definition, \(C \in Descendants (C)\) by convention.

  17. The paper by Pearl and Dechter (1996) extends previous results by Spirtes (1995) and Koster (1996) who have shown that the \(d\)-separation test is valid for cyclic graphs with linear equations and normal distributions over the error terms. Given that we restrict attention to discrete variables in this paper (see Sect. 4), we will only discuss Pearl and Dechter.

  18. Disturbances are not explicitly mentioned in Sect. 4, but they can easily be introduced. Disturbances are variables that represent errors due to omitted factors (see Pearl 2000, p. 27). The assumption of uncorrelated disturbances is not a severe restriction. Given a graph \(G\) and associated probability distribution \(P\), such that not all disturbances are independent, one may construct an augmented graph \(G'\) in which all disturbances are independent. The augmented graph \(G'\) is obtained by adding, for each pair of dependent disturbances, a dummy root node as a common cause of the disturbances (Pearl and Dechter 1996, p. 422).

  19. Unlike in the acyclic case, however, ‘the joint distribution of feedback systems cannot be written as a product of the conditional distributions of each child variable, given its parents’ (Pearl and Dechter 1996, p. 420). Hence factorization, on which we relied in Sect. 4, cannot be applied to DCGs. This can be shown by means of a simple example by (Spirtes et al. (2000), §12.1.2). Applying the factorization to the graph \(X \leftrightarrows Y\), would lead to \(P(X, Y) = P(X \mid Y) \times P(Y \mid X)\), which would mean that \(X\) and \(Y\) are independent, contrary to what the graph suggests.

  20. Simultaneous causation and backwards causation do not fit this picture. However, such cases rarely if ever occur in models of mechanisms, so we set them aside in this paper.

  21. This procedure is outlined by Lauritzen.et.al (1990). Their alternative test for \(d\)-separation is equivalent to the one specified in Sect. 5.1 and is used by Pearl and Dechter (1996) in their discussion of \(d\)-separation for DCGs.

  22. Separation in undirected graphs (such as moral graphs) is defined as follows: let \(G\) be an undirected graph with vertex set \(V\), then two sets of vertices \(X, Y \subseteq V\) are separated by \(Z \subseteq V\) if and only if every path (sequence of undirected edges) from each vertex in \(X\) to each vertex in \(Y\) contains some vertex in \(Z\).

  23. Prima facie, this approach runs counter to Bechtel’s appeal not to model mechanisms sequentially (see Sect. 2). By means of the transition network, however, the cyclic organization is captured as well. We would like to thank Michael Wilde for pointing out this seeming incongruity.

  24. To take a concrete example, suppose that variables \(A\) and \(B\) directly cause \(C\) which directly causes \(D\) which in turn directly causes each of \(A\) and \(B\). A causal DAG at time \(0\) obtained by unwinding this cycle might have arrows from \(A_0\) and \(B_0\) to \(C_0\) and an arrow from \(C_0\) to \(D_0\). The Markov condition requires that \(A_0\) and \(B_0\) be probabilistically independent. But these two variables have a common cause not represented in the graph—the previous instance of \(D\)—which can render them probabilistically dependent. Thus the Markov condition can fail in this causal graph.

  25. For a detailed account of mutual manipulability, see (Craver (2007), pp. 152–160). For a recent critical discussion of Craver’s claim that interlevel constitutive relations cannot be causal, and whether this claim is compatible with his mutual manipulability account of constitutive relevance, see Leuridan (2012).

  26. See, among others, Hausman and Woodward (1999, 2004a, b), Cartwright (2001, 2002), Williamson (2005) and Steel (2006).

References

  • Bechtel, W. (2011). Mechanism and biological explanation. Philosophy of Science, 78(4), 533–557.

    Article  Google Scholar 

  • Bechtel, W., & Abrahamsen, A. (2005). Explanation: A mechanist alternative. Studies in History and Philosophy of Biological and Biomedical Sciences, 36(2), 421–441.

    Google Scholar 

  • Bernard, A., & Hartemink, A. (2005). Informative structure priors: Joint learning of dynamic regulatory networks from multiple types of data. In R. Altman, A. K. Dunker, L. Hunter, T. Jung, & T. Klein (Eds.), Proceedings of the Pacific symposium on biocomputing (PSB05) (pp. 459–470). Hackensack, NJ: World Scientific.

  • Boon, N. A., Colledge, N. R., Walker, B. R., & Hunter, J. A. (Eds.). (2006). Davidson’s principles & practice of medicine (20th ed.). Edinburgh: Churchill Livingstone.

  • Bouchaffra, D. (2010). Topological dynamic Bayesian networks. In Proceedings of the twentieth IEEE international conference on pattern recognition (pp. 898–901).

  • Cartwright, N. (2001). What is wrong with Bayes nets? The Monist, 84(2), 242–264.

    Article  Google Scholar 

  • Cartwright, N. (2002). Against modularity, the causal Markov condition, and any link between the two: Comments on Hausman and Woodward. The British Journal for the Philosophy of Science, 53(3), 411–453.

    Google Scholar 

  • Casini, L., Illari, P. M., Russo, F., & Williamson, J. (2011). Models for prediction, explanation and control: Recursive Bayesian networks. Theoria, 26(1), 5–33.

    Google Scholar 

  • Cowell, R. G., Dawid, A. P., Lauritzen, S. L., & Spiegelhalter, D. J. (1999). Probabilistic networks and expert systems. Berlin: Springer.

    Google Scholar 

  • Craver, C. F. (2006). When mechanistic models explain. Synthese, 153(3), 355–376.

    Article  Google Scholar 

  • Craver, C. F. (2007). Explaining the brain: Mechanisms and the mosaic unity of neuroscience. Oxford: Clarendon Press.

    Book  Google Scholar 

  • Crunelli, V., & Hughes, S. W. (2010). The slow (\({<}\) 1 Hz) rhythm of non-REM sleep: A dialogue between three cardinal oscillators. Nature Neurosciences, 13(1), 9–17.

    Article  Google Scholar 

  • Dean, T., & Kanazawa, K. (1989). A model for reasoning about persistence and causation. Computational Intelligence, 5(3), 142–150.

    Article  Google Scholar 

  • Doshi-Velez, F., Wingate, D., Tenenbaum, J., & Roy, N. (2011). Infinite dynamic bayesian networks. In L. Getoor & T. Scheffer (Eds.), Proceedings of the 28th international conference on machine learning (ICML) (pp. 913–920). Montreal: Omnipress.

  • Friedman, N., Murphy, K. P., & Russell, S. J. (1998). Learning the structure of dynamic probabilistic networks. In G. F. Cooper & S. Moral (Eds.), Proceedings of the fourteenth conference on uncertainty in artificial intelligence (UAI) (pp. 139–147), San Mateo, CA. Morgan Kaufmann.

  • Gebharter, A., & Kaiser, M. I. (2012). Causal graphs and mechanisms. In A. Hüttemann, M. I. Kaiser, & O. Scholz (Eds.), Explanation in the special sciences. The case of biology and history. Synthese Library. Dordrecht: Springer.

  • Ghahramani, Z. (1998). Learning dynamic bayesian networks. In C. Giles & M. Gori (Eds.), Adaptive processing of sequences and data structures. Lecture notes in computer science (Vol. 1387, pp. 168–197). Berlin: Springer. doi:10.1007/BFb0053999.

  • Glennan, S. S. (1996). Mechanisms and the nature of causation. Erkenntnis, 44(1), 49–71.

    Article  Google Scholar 

  • Glennan, S. S. (2002). Rethinking mechanistic explanation. Philosophy of Science, 69(3), S342–S353.

    Article  Google Scholar 

  • Glennan, S. S. (2005). Modeling mechanisms. Studies in History and Philosophy of Biological and Biomedical Sciences, 36(2), 443–464.

    Google Scholar 

  • Gottlieb, D. J., Yenokyan, G., Newman, A. B., O’Connor, G. T., Punjabi, N. M., Quan, S. F., et al. (2010). Prospective study of obstructive sleep apnea and incident coronary heart disease and heart failure. Circulation, 122(4), 352–360.

    Article  Google Scholar 

  • Grandner, M., Hale, L., Moore, M., & Patel, N. V. (2010). Mortality associated with short sleep duration: The evidence, the possible mechanisms, and the future. Sleep Medicine Reviews, 14(3), 191–203.

    Article  Google Scholar 

  • Hausman, D. M., & Woodward, J. (1999). Independence, invariance and the causal Markov condition. The British Journal for the Philosophy of Science, 50(4), 521–583.

    Article  Google Scholar 

  • Hausman, D. M., & Woodward, J. (2004a). Manipulation and the causal Markov condition. Philosophy of Science, 55(5), 147–161.

    Google Scholar 

  • Hausman, D. M., & Woodward, J. (2004b). Modularity and the causal Markov condition: A restatement. The British Journal for the Philosophy of Science, 55(1), 147–161.

    Article  Google Scholar 

  • Koster, J. T. A. (1996). Markov properties of nonrecursive causal models. Annals of Statistics, 24(5), 2148–2177.

    Google Scholar 

  • Lauritzen, S., Dawid, A., Larsen, B., & Leimer, H.-G. (1990). Independence properties of directed Markov fields. Networks, 20(5), 491–505.

    Article  Google Scholar 

  • Lazebnik, Y. (2002). Can a biologist fix a radio?—Or, what I learned while studying apoptosis. Cancer Cell, 2, 179–182.

    Article  Google Scholar 

  • Leuridan, B. (2010). Can mechanisms really replace laws of nature? Philosophy of Science, 77(3), 317–340.

    Article  Google Scholar 

  • Leuridan, B. (2012). Three problems for the mutual manipulability account of constitutive relevance in mechanisms. The British Journal for the Philosophy of Science, 63(2), 399–427.

    Article  Google Scholar 

  • Leuridan, B. (2014). The structure of scientific theories, explanation, and unification. A causal-structural account. The British Journal for the Philosophy of Science doi:10.1093/bjps/axt015.

  • Machamer, P., Darden, L., & Craver, C. F. (2000). Thinking about mechanisms. Philosophy of Science, 67(1), 1–25.

    Article  Google Scholar 

  • Marshall, L., Helgadóttir, H., Mölle, M., & Born, J. (2006). Boosting slow oscilllations during sleep potentiates memory. Nature, 444(7119), 610–613.

    Article  Google Scholar 

  • McNicholas, W., & Bonsignore, M. (2007). Sleep apnoea as an independent risk factor for cardiovascular disease: Current evidence, basic mechanisms and research priorities. European Respiratory Journal, 29(1), 156–178.

    Article  Google Scholar 

  • Mitchell, S. (2009). Unsimple truths: Science, complexity, and policy. Chicago, IL: University of Chicago Press.

    Book  Google Scholar 

  • Murphy, K. P. (2002). Dynamic Bayesian networks: Representation, inference and learning. PhD thesis, Computer Science, University of California, Berkeley.

  • Neal, R. (2000).On deducing conditional independence from \(d\)-separation in causal graphs with feedback: The uniqueness condition is not suffient. Journal of Artificial Intelligence Research, 12, 87–91.

    Google Scholar 

  • Neapolitan, R. E. (1990). Probabilistic reasoning in expert systems: Theory and algorithms. New York: Wiley.

    Google Scholar 

  • Nervi, M. (2010). Mechanisms, malfunctions and explanation in medicine. Biology and Philosophy, 25(2), 215–228.

    Article  Google Scholar 

  • Pearl, J. (1988). Probabilistic reasoning in intelligent systems: Networks of plausible inference. San Mateo, CA: Morgan Kaufmann.

    Google Scholar 

  • Pearl, J. (2000). Causality: Models, reasoning, and inference. Cambridge: Cambridge University Press.

    Google Scholar 

  • Pearl, J., & Dechter, R. (1996). Identifying independencies in causal graphs with feedback. In Uncertainty in artificial intelligence: Proceedings of the twelfth conference (pp. 420–426). San Mateo, CA. Morgan Kaufmann.

  • Perini, L. (2005a). Explanation in two dimensions: Diagrams and biological explanation. Biology and Philosophy, 20(2–3), 257–269.

    Article  Google Scholar 

  • Perini, L. (2005b). The truth in pictures. Philosophy of Science, 72(1), 262–285.

    Article  Google Scholar 

  • Perini, L. (2005c). Visual representations and confirmation. Philosophy of Science, 72(5), 913–926.

    Article  Google Scholar 

  • Seidenfeld, T. (1987). Entropy and uncertainty. In I. B. MacNeill & G. J. Umphrey (Eds.), Foundations of statistical inference (pp. 259–287). Dordrecht: Reidel.

  • Spirtes, P. (1995). Directed cyclic graphical representation of feedback models. In P. Besnard & S. Hanks (Eds.), Proceedings of the eleventh conference on uncertainty in artificial intelligence (pp. 491–498). San Mateo, CA: Morgan Kaufmann.

    Google Scholar 

  • Spirtes, P., Glymour, C., & Scheines, R. (2000). Causation, prediction, and search. Cambridge, MA: MIT Press.

    Google Scholar 

  • Spirtes, P., Glymour, C., Scheines, R., & Tillman, R. (2010). Automated search for causal relations—Theory and practice. In R. Dechter, H. Geffner, & J. Y. Halpern (Eds.), Heuristics, probability and causality: A tribute to Judea Pearl (pp. 467–506). London: College Publications.

    Google Scholar 

  • Steel, D. (2006). Comment on Hausman & Woodward on the causal Markov condition. British Journal for the Philosophy of Science, 57(1), 219–231.

    Article  Google Scholar 

  • Williamson, J. (2005). Bayesian nets and causality: Philosophical and computational foundations. Oxford: Oxford University Press.

  • Williamson, J. (2010). In defence of objective Bayesianism. Oxford: Oxford University Press.

    Book  Google Scholar 

  • Woodward, J. (2002). What is a mechanism? A counterfactual account. Philosophy of Science, 69(3), S366–S377.

    Article  Google Scholar 

  • Woodward, J. (2003). Making things happen. A theory of causal explanation. New York: Oxford University Press.

  • Yumino, D., Redolfi, S., Ruttanaumpawan, P., Su, M., Smith, S., Newton, G. E., et al. (2010). Nocturnal rostral fluid shift. Circulation, 121(14), 1598–1605.

    Article  Google Scholar 

Download references

Acknowledgments

We would like to thank Lorenzo Casini, George Darby, Phyllis Illari, Mike Joffe, Federica Russo, and Michael Wilde for their helpful comments on this paper. Jon Williamson’s research is supported by the UK Arts and Humanities Research Council. Bert Leuridan is Postdoctoral Fellow of the Research Foundation—Flanders (FWO).

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Bert Leuridan.

Additional information

The authors are listed alphabetically.

Appendix: Transforming a moral graph into a DAG

Appendix: Transforming a moral graph into a DAG

Here we present the algorithm of (Williamson (2005), §5.7) for transforming an undirected graph \({\mathcal {G}}\) into a directed acyclic graph \({\mathcal {H}}\) which preserves the required independencies (Williamson 2005, Theorem 5.3): if \(Z\) \(d\)-separates \(X\) from \(Y\) in the DAG \({\mathcal {H}}\) then \(X\) and \(Y\) are separated by \(Z\) in the undirected graph \({\mathcal {G}}\); this separation in \({\mathcal {G}}\) implies that \(X\) and \(Y\) are probabilistically independent conditional on \(Z\); hence, \(d\)-separation in \({\mathcal {H}}\) implies that \(X\) and \(Y\) are probabilistically independent conditional on \(Z\). Thus \({\mathcal {H}}\) can be used as the graph of a Bayesian network.

An undirected graph is triangulated if for every cycle involving four or more vertices there is an edge in the graph between two vertices that are non-adjacent in the cycle. The first step of the procedure is to construct a triangulated graph \({\mathcal {G}}^T\) from the undirected graph \({\mathcal {G}}\). One of a number of standard triangulation algorithms can be applied to construct \({\mathcal {G}}^T\) (see, e.g., Neapolitan 1990, §3.2.3; Cowell et al. 1999, §4.4.1).

Next, re-order the variables in \(V\) according to maximum cardinality search with respect to \({\mathcal {G}}^T\): choose an arbitrary vertex as \(V_{1}\); at each step select the vertex which is adjacent to the largest number of previously numbered vertices, breaking ties arbitrarily. Let \(D_{1},\ldots ,D_{l}\) be the cliques (i.e., maximal complete subgraphs) of \({\mathcal {G}}^T\), ordered according to highest labelled vertex. Let \(E_{j} = D_{j}\cap (\bigcup _{i=1}^{j-1} D_{i})\) and \(F_{j} = D_{j} {\setminus } E_{j}\), for \(j=1,\ldots ,l\).

Finally, construct a DAG \({\mathcal {H}}\) as follows. Take variables in \(V\) as vertices. Step 1: add an arrow from each vertex in \(E_{j}\) to each vertex in \(F_{j}\), for \(j=1,\ldots ,l\). Step 2: add further arrows to ensure that there is an arrow between each pair of vertices in \(D_{j}, j=1,\ldots ,l\), taking care that no cycles are introduced (there is always some orientation of an added arrow which will not yield a cycle).

Rights and permissions

Reprints and permissions

About this article

Cite this article

Clarke, B., Leuridan, B. & Williamson, J. Modelling mechanisms with causal cycles. Synthese 191, 1651–1681 (2014). https://doi.org/10.1007/s11229-013-0360-7

Download citation

  • Received:

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s11229-013-0360-7

Keywords

Navigation