Skip to main content

Advertisement

Log in

Rethinking explainability: toward a postphenomenology of black-box artificial intelligence in medicine

  • Original Paper
  • Published:
Ethics and Information Technology Aims and scope Submit manuscript

Abstract

In recent years, increasingly advanced artificial intelligence (AI), and in particular machine learning, has shown great promise as a tool in various healthcare contexts. Yet as machine learning in medicine has become more useful and more widely adopted, concerns have arisen about the “black-box” nature of some of these AI models, or the inability to understand—and explain—the inner workings of the technology. Some critics argue that AI algorithms must be explainable to be responsibly used in the clinical encounter, while supporters of AI dismiss the importance of explainability and instead highlight the many benefits the application of this technology could have for medicine. However, this dichotomy fails to consider the particular ways in which machine learning technologies mediate relations in the clinical encounter, and in doing so, makes explainability more of a problem than it actually is. We argue that postphenomenology is a highly useful theoretical lens through which to examine black-box AI, because it helps us better understand the particular mediating effects this type of technology brings to clinical encounters and moves beyond the explainability stalemate. Using a postphenomenological approach, we argue that explainability is more of a concern for physicians than it is for patients, and that a lack of explainability does not introduce a novel concern to the physician–patient encounter. Explainability is just one feature of technological mediation and need not be the central concern on which the use of black-box AI hinges.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Similar content being viewed by others

Notes

  1. In this paper, we focus our analysis mainly on physician use of black-box AI, as they most often deploy AI models and use particular models for diagnosis and/or treatment. Yet we recognize that other clinicians may also grapple with the use of AI in their work, and thus our analysis can be applied to healthcare workers in other roles, as well.

  2. While we agree with much of Kiran’s analysis, it is worth questioning whether this aspect of his conclusion is justified. If technology completely frames and constrains what we see, like Kiran says, are we free to choose the extent to which we are constrained? It would distract from our thesis to address this question in full here, but we worry that Kiran’s suggestion does not take seriously the extent to which technology determines our gaze.

  3. This can lead to widespread concerns about bias and discrimination, discussed further below.

  4. Even though patients do not often request a robust explanation of the medical technologies routinely used in their care, there remains a normative question of whether they should, especially given certain features, such as potential algorithmic bias. While this is an interesting question worthy of further study, inquiring into the individual responsibilities of patients is beyond the scope of this paper.

  5. This also raises legal and ethical concerns of accountability and liability at the level of the healthcare system. Who is to blame when an algorithm turns out to be flawed or systematically prescribes harmful treatments: the individual physician? The hospital? The firm that designed the AI model? (Grote & Berens, 2020).

  6. While Ihde also introduces the epistemological magnification-reduction effects of technological mediation, we will focus here on Kiran’s three dimensions of technological mediation (enabling-constraining, revealing-concealing, involving-alienating), as these sufficiently build from Ihde and include elements of magnification-reduction throughout. In general, black box AI magnifies some parameters and necessarily reduces others, but the specific data points being magnified and reduced are context-specific and thus warrant unique analysis for individual AI algorithms.

  7. This also highlights the multistability of black-box AI, or the various ways in which different users can engage with the technology. What feels very important to physicians may not actually be important to patients, and vice versa.

References

  • Aho, K. (2018). Existential medicine: Essays on health and illness. Rowman & Littlefield International.

    Google Scholar 

  • Angwin, J., Larson, J., Mattu, S., & Kirchner, L. (2016). Machine bias. ProPublica. www.propublica.org/article/machine-bias-risk-assessments-in-criminal-sentencing

  • Bishop, J. P. (2011). The anticipatory corpse: Medicine, power, and the care of the dying. University of Notre Dame Press.

    Book  Google Scholar 

  • Brouillette, M. (2017). Deep learning is a black box, but health care won’t mind. MIT Technology Review. https://www.technologyreview.com/2017/04/27/242905/deep-learning-is-a-black-box-but-health-care-wont-mind/

  • Campolo, A., Sanfilippo, M., Whittaker, M., & Crawford, M. (2017). AI Now 2017 Report. AI Now Institute. https://ainowinstitute.org/AI_Now_2017_Report.pdf

  • Clarke, A. 1973. Profiles of the future: An inquiry into the limits of the possible. Popular Library.

  • Gadamer, H.-G. (1996). The enigma of health: The art of healing in a scientific age. Stanford University Press.

    Google Scholar 

  • Gertz, N. (2018). Nihilism and technology. Rowman & Littlefield.

    Google Scholar 

  • Goodman, K. W. (2007). Ethical and legal issues in decision support. In E. S. Berner (Ed.), Clinical decision support systems: Theory and Practice (2nd ed., pp. 126–39). Springer.

    Chapter  Google Scholar 

  • Grote, T. & Berens, P. (2020). On the ethics of algorithmic decision-making in healthcare. Journal of Medical Ethics, 46, 205–211.

  • Hadler, R. A., Clapp, J. T., Chung, J. J., Gutsche, J. T., & Fleisher, L. A. (2021). Escalation and withdrawal of treatment for patients on extracorporeal membrane oxygenation (ECMO): A qualitative study. Annals of Surgery. https://doi.org/10.1097/SLA.0000000000004838

    Article  Google Scholar 

  • Hannun, A. Y., Rajpurkar, P., Haghpanahi, M., et al. (2019). Cardiologist-level arrhythmia detection and classification in ambulatory electrocardiograms using a deep neural network. Nature Medicine, 25, 65–69. https://doi.org/10.1038/s41591-018-0268-3

    Article  Google Scholar 

  • Hawley, K. (2015). Trust and distrust between patient and doctor. Journal of Evaluation in Clinical Practice, 21(5), 798–801.

    Article  Google Scholar 

  • Heidegger, M. (1977). The question concerning technology, and other essays. Harper and Row.

    Google Scholar 

  • Heidegger, M., & Boss, M. (2001). Zollikon seminars: Protocols, conversations, letters. Northwestern University Press.

    Google Scholar 

  • Hofmann, B., & Svenaeus, F. (2018). How medical technologies shape the experience of illness. Life Sciences, Society and Policy, 14, 3.

    Article  Google Scholar 

  • Holzinger, A., Carrington, A., & Müller, H. (2020). Measuring the quality of explanations: The system causability scale (SCS). KI - Künstliche Intelligenz, 34(2), 193–198. https://doi.org/10.1007/s13218-020-00636-z

    Article  Google Scholar 

  • Ihde, D. (1977). Experimental phenomenology: An introduction. Putnam.

    Google Scholar 

  • Ihde, D. (1990). Technology and the lifeworld: From garden to earth. Indiana University Press.

    Google Scholar 

  • Ihde, D. (2002). Bodies in technology. University of Minnesota Press.

    Google Scholar 

  • Kiran, A. H. (2015). Four dimensions of technological mediation. In R. Rosenberger & P.-P. Verbeek (Eds.), Postphenomenological investigations: Essays on Human-Technology Relations (pp. 123–140). Lexington Book.

    Google Scholar 

  • Lipton, C. Z. (2016). The mythos of model interpretability. arXiv preprint. https://arxiv.org/pdf/1606.03490.pdf.

  • London, A.J. (2018). Groundhog day for medical artificial intelligence. Hastings Center Report, 48(3).

  • London, A. J. (2019). Artificial intelligence and black-box medical decisions: Accuracy versus explainability. Hastings Center Report, 49(1), 15–21.

    Article  Google Scholar 

  • Malone, J. (2019). Invasive medical technology: A postphenomenological variational analysis (Publication No. 13881260) [Doctoral Dissertation, Saint Louis University]. ProQuest Dissertations Publishing.

  • McDougall, R. J. (2019). Computer knows best? The need for value-flexibility in medical AI. Journal of Medical Ethics, 45(3), 156–160.

    Article  MathSciNet  Google Scholar 

  • Mukherjee, S. (2017). A.I. versus M.D. The New Yorker. https://www.newyorker.com/magazine/2017/04/03/ai-versus-md

  • Nemati, S., Holder, A., Razmi, F., et al. (2018). An interpretable machine learning model for accurate prediction of sepsis in the ICU. Critical Care Medicine, 46, 547–553. https://doi.org/10.1097/CCM.0000000000002936

    Article  Google Scholar 

  • O’Neil, C. (2016). Weapons of math destruction: How big data increases inequality and threatens democracy. Crown.

    MATH  Google Scholar 

  • Oudshoorn, N. E. J. (2011). Telecare technologies and the transformation of healthcare. Palgrave Macmillan.

    Book  Google Scholar 

  • Rajkomar, A., Oren, E., Chen, K., et al. (2018). Scalable and accurate deep learning with electronic health records. NPJ Digital Medicine, 1, 18. https://doi.org/10.1038/s41746-018-0029-1

    Article  Google Scholar 

  • Rosenberger, R., & Verbeek, P.-P. (2015). Postphenomenological investigations: Essays on human-technology relations. Lexington Books.

    Google Scholar 

  • Simonite, T. (2018). Google's AI guru wants computers to think more like brains. Wired. www.wired.com/story/googles-ai-guru-computers-think-more-like-brains/

  • Szegedy, C., Vanhoucke, V., Ioffe, S., et al. (2016). Rethinking the inception architecture for computer vision. In Proceedings of the 29th IEEE Conference on Computer Vision and Pattern Recognition. Las Vegas, Nevada, 26 June–1 July 2016. Piscataway, NJ, pp. 2818–26.

  • Topol, E. J. (2019). Deep medicine: How artificial intelligence can make healthcare human again. Basic Books.

    Google Scholar 

  • Verbeek, P.-P. (2005). What things do: Philosophical reflections on Technology, Agency, and Design. Pennsylvania State University Press.

    Book  Google Scholar 

  • Wang, F., Kaushal, R., & Khullar, D. (2020). Should health care demand interpretable artificial intelligence or accept “black box” medicine? Annals of Internal Medicine, 172, 59–60. https://doi.org/10.7326/M19-2548

    Article  Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Annie B. Friedrich.

Additional information

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Friedrich, A.B., Mason, J. & Malone, J.R. Rethinking explainability: toward a postphenomenology of black-box artificial intelligence in medicine. Ethics Inf Technol 24, 8 (2022). https://doi.org/10.1007/s10676-022-09631-4

Download citation

  • Accepted:

  • Published:

  • DOI: https://doi.org/10.1007/s10676-022-09631-4

Keywords

Navigation