Skip to main content
Log in

Explanations in AI as Claims of Tacit Knowledge

  • S.I. : Machine learning: Prediction Without Explanation?
  • Published:
Minds and Machines Aims and scope Submit manuscript

Abstract

As AI systems become increasingly complex it may become unclear, even to the designer of a system, why exactly a system does what it does. This leads to a lack of trust in AI systems. To solve this, the field of explainable AI has been working on ways to produce explanations of these systems’ behaviors. Many methods in explainable AI, such as LIME (Ribeiro et al. in Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, 2016), offer only a statistical argument for the validity of their explanations. However, some methods instead study the internal structure of the system and try to find components which can be assigned an interpretation. I believe that these methods provide more valuable explanations than those statistical in nature. I will try to identify which explanations can be considered internal to the system using the Chomskyan notion of tacit knowledge. I argue that each explanation expresses a rule, and through the localization of this rule in the system internals, we can take a system to have tacit knowledge of the rule. I conclude that the only methods which are able to sufficiently establish this tacit knowledge are those along the lines of Olah (Distill 2(11): 4901–4911, 2017), and therefore they provide explanations with unique strengths.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Institutional subscriptions

Similar content being viewed by others

Data Availability

None.

Code Availability

None.

Notes

  1. In AI, the terms ‘decision’, ‘action’, ‘output’ and ‘result’ are generally used interchangeably. From an agnostic information processing perspective, these are all just the signal(s) the system outputs to the environment, but depending on the context one of the terms may be more appropriate than the others. If the signal is used to move a robot arm, it can be described as an action, if it is used to inform another action, it may be called a decision. I prefer generic terms like ‘output’ when discussing the more theoretical aspects and terms like ‘decision’ when discussing practical implications, to emphasize the consequences these outputs may have when viewing the systems as actors in a societal context.

  2. By behavior, I mean the outputs of an AI system, or more specifically: the relationships between inputs and the corresponding outputs that the system produces. How the system arrives at these outputs is not a part of the behavior: it is entirely treated as a black box.

  3. For example via a saliency map, see Sect. 6 for details.

  4. Here the term ‘knowledge’ is used a bit more loosely than it generally is in epistemology (e.g. it does not need to be true). Knowledge is roughly meant to mean ‘information an agent uses to determine their actions’. This usage is common in psychological and AI literature. It can alternatively be read as ‘putative knowledge’ or even just ‘information’.

  5. This claim is taken from an example in Miller (2019, Fig. 2) where it is used to illustrate how an agent might explain the results of a system which classifies arthropods.

  6. As an example of this terminology, Blum and Langley (1997) named the “selection of relevant features” as one of the central problems in machine learning, and describes the way in which a machine learning program can learn about concepts in terms of features: “[Concept learning is] deciding which features to use in describing the concept and deciding how to combine those features”.

  7. In other words, I assume that the explanation ‘y because p’ is complete, in the sense that it is not needed to take into account any factors other than p to determine that y. If we instead assume p is (only) a necessary condition for y, we obtain a counterfactual theory of causation (Menzies and Beebee 2020): if p had not been there, y would not have been the case.

  8. Note that ‘computation’ and ‘behavior’ are two separate things: while the behavior is only the observed input-output relationship, by computation I am referring to all the steps in between that lead the system to produce this output. While looking at the behavior treats the system as a black box, to get a complete picture of the computation we in fact need to have full access to the system’s internals.

  9. The absence of theoretical structure is a consequence of the black-boxness of the model. In machine learning, we start with a high-level “task analysis” (Clark 1990, p. 215), and use this to construct an implementation through some process of trial and error. But we are lacking what Clark calls a “fully articulated competence theory” of the task, one expressed in traditional mathematical language and incorporating symbols which are meaningful to us.

  10. Many linguists will disagree, and might rather speak of a statistical kind of knowledge. They will argue that simply the fact that the order ‘eats an apple’ is much more likely to occur in conversation than ‘an apple eats’ allows one to judge the first sentence as correct, without any internalization of an abstract concept such as ‘word order’. This is an application of the so-called distributional hypothesis (Harris 1970). However, this statistical view is exactly the prevalent view in machine learning, to which I am trying to provide a complementary perspective, so I will not take time to discuss it here.

  11. The statement ‘\(p \Rightarrow q\)’ is equivalent to saying that p is a sufficient condition for q. If we would adhere to a counterfactual theory of explanation, where p is (only) a necessary condition for q, we would instead have that \(q \Rightarrow p\).

  12. It should be noted that the kind of tacit knowledge we are looking for here is knowledge that is not explicit, but can be made explicit, as it is represented in explicit form by the claims we eventually make about the system. Collins (2010) refers to this as explicable knowledge, and gives a taxonomy of tacit knowledge, in which he orders different types by “the degree of resistance of the tacit knowledge to being made explicit” (Collins 2010, p. 85). In his terminology, the knowledge we are looking for might be relational tacit knowledge, or possibly somatic tacit knowledge.

  13. At most, there may be an approximate representation of R, but its existence will in any case be dependent on our choice to interpret it as such.

  14. One could say that by ‘discovering’ the rule we are in effect performing the event of formulation, and turning it from a tacit rule to an explicit one. This is true in some sense, as we are creating an explicit formulation of the rule. However, the rule itself remains tacit, as the way in which it is possessed by the agent (i.e. the internal representation of the rule) does not change. It remains distinct from an explicit rule in that the agent themselves has still not formulated it. The explicit version of the rule is only a result of our interpretation, and so the agent is not beholden to it. Any ‘meaning’ or semantic interpretation imparted onto the processing performed by the agent using this statement of the rule will be our responsibility as well. As such, making tacit knowledge in an agent explicit does not change the nature of the rule, it only helps to verify our interpretation of the agent’s behavior.

  15. It should be noted that it is not entirely evident why exactly Chomsky’s and Davies’ ‘tacit knowledge’ should both be referring to the same kind of knowledge. Instead, it is better to view Davies’ definition as a separate kind of knowledge, with the same spirit as Chomsky’s notion: to provide a meaningful middle ground where one can speak of knowledge that an agent possesses without having to either admit all possible knowledge that is compatible with the behavior, or exclude all knowledge whose presence cannot be directly observed.

  16. One can require various levels of strictness in this localization: must there be a specific (combination of) processing unit(s) in the system which represents this concept, in the same way, 100% of the time? Or is it only a requirement that we always need to be able to obtain an answer on the ‘8 legs’-question from ‘somewhere’, but that its representation may change from case to case, as long as we ‘know where to look’? In other words, do we only require that T is expressible as some mathematical function, or do we place further restrictions on it?

    I am inclined to not be too strict in this regard, if only for the fact that even a physical implementation of a straightforward symbolic program may have various physical (or implementational) states corresponding to the same symbolic state in different contexts, for example for purposes of making the computation more efficient. Clark (1990, p. 207) also seems to hint at this in his assessment of Davies’ work, saying that the representative state in the system “is not a simple physical state so must as a state of the virtual machine over which the processing story is defined”. In addition, if we restrict the definition of a ‘component’ too much it will be very hard to deal with so-called distributed representations of concepts. Schröder (1998) notices this, and argues we should instead allow for several common causal factors in Davies’ definition, which is similar to allowing a rule to have a representation which varies between situations.

    However, having too loose a requirement clearly devalues the term ‘localizable’ to be nearly meaningless: if the representation of ‘having 8 legs’ is different in every scenario, the argument for presence of ‘having 8 legs’ in the system seems quite weak. Presumably, we can at least speak of a spectrum: the more localizable a rule is, the stronger the claim of its presence will be.

  17. This same question is also being tackled by other fields, such as neuro-symbolic integration (also called hybrid learning).

  18. This example is again taken from Miller (2019, Fig. 2).

  19. In more recent research, the authors of LIME have also worked on methods that cast the relationship between input and output into the form of rules. These rules specify a sufficient condition for a certain output to be produced (Ribeiro et al. 2018), which means they correspond to the rules discussed in this paper. While their rule-based method does make it more feasible to produce the kind of explanations I am after, it is still susceptible to the criticism that it does not look at the internal state of the system, and is therefore only descriptive of the behavior.

  20. For a counterfactual theory of causation, it also cannot establish whether the feature is a necessary condition (e.g. to be considered a spider, an animal needs to have 8 legs). In this case the problem is that the actual relationship may vary greatly across different inputs, so that there may be, for example, an unobserved input that is still considered a spider but is not considered to have 8 legs.

  21. His go-to example is the cluster analysis of NETtalk, a connectionist network which converts English words into (textually encoded) phonemes (Sejnowski and Rosenberg 1987). By performing a clustering procedure over the internal states of the system, the system could be found to have rediscovered (among other things) that there is a useful division to be made between vowels and consonants (Clark 1993, pp. 52-54). Cluster analysis seems to have fallen by the wayside, but it is similar to modern day applications of dimensionality reduction methods to discover ‘typical’ states of neural networks (e.g. Li et al. 2020). It has also influenced the method of representational similarity analysis (RSA) (Kriegeskorte 2008), which is used to perform experiments concerning the ways in which the human brain represents information.

  22. Note that, while the computation a single neuron performs is very simple (a nearly linear function of its inputs), if we consider a neuron a few layers deep in the network it will be provided with the results of all prior processing, and so the output of this neuron is in fact a very complex function of the inputs. Therefore inputs which maximize it may contain many kinds of structure, possibly encoding high-level concepts.

  23. ‘Visualization’ implies a visual representation of the feature, which is often appropriate because many black-box systems of interest deal with image processing. However, what seems more important to me is the method by which a visualization of a feature is obtained, i.e. the optimization part. That the feature is then presented in visual form seems irrelevant, it might just as well be an representative piece of text or an audio sample depending on the system.

  24. This relates back to the question about the ‘strictness’ of the localization we require: see footnote 16. While it would be nice if a single concept is represented by a single neuron, because that is highly localizable, it might not realistically happen very often and we might encounter mostly polysemantic neurons instead. However, it might still be the case that a certain combination of these neurons responds to a specific concept (and another combination to another concept, etc.): this is the idea of distributed representation. For example, in representational similarity analysis (Kriegeskorte 2008) correlations between various internal states corresponding to similar concepts are used to justify that the concepts are being represented, without having to unambiguously localize the concept. Basically, to find a distributed representation of a concept we should be able to perform some transformation on the combined state of several neurons in order to demonstrate its presence. This raises some further philosophical questions, however, such as whether or not the concept is not partially (or even mostly) represented by the transformation we apply (i.e. our interpretation), instead of in the system itself. Just as in the previous discussion, if we allow for some nuance we can at least argue for a correlation between the two: “the more localizable a rule is, the stronger the claim of its presence will be”.

  25. In practice, there may of course be further slight numerical approximations, such as neglecting a causal path containing near zero weights. But these are easily justifiable. I am mostly concerned with approximations which are hard to justify, such as the proper interpretation of a fuzzy concept such as ‘blue’.

  26. The journal Distill publishes many of these new approaches to visualizing or otherwise making explanations about neural networks accessible, for example using interactive visual maps (Carter et al. 2019) or by annotating sentences with the influences between characters when performing natural language prediction (Madsen 2019).

References

Download references

Funding

None.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Nardi Lam.

Ethics declarations

Conflict of interest

None

Additional information

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Lam, N. Explanations in AI as Claims of Tacit Knowledge. Minds & Machines 32, 135–158 (2022). https://doi.org/10.1007/s11023-021-09588-1

Download citation

  • Received:

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s11023-021-09588-1

Keywords

Navigation