Accuracy and Interpretability: Struggling with the Epistemic Foundations of Machine Learning-Generated Medical Information and Their Practical Implications for the Doctor-Patient Relationship
Tóm tắt
The initial successes in recent years in harnessing machine learning (ML) technologies to improve medical practice and benefit patients have attracted attention in a wide range of healthcare fields. Particularly, it should be achieved by providing automated decision recommendations to the treating clinician. Some hopes placed in such ML-based systems for healthcare, however, seem to be unwarranted, at least partially because of their inherent lack of transparency, although their results seem convincing in accuracy and reliability. Skepticism arises when the physician as the agent responsible for the implementation of diagnosis, therapy, and care is unable to access the generation of findings and recommendations. There is widespread agreement that, generally, a complete traceability is preferable to opaque recommendations; however, there are differences about addressing ML-based systems whose functioning seems to remain opaque to some degree—even if so-called explicable or interpretable systems gain increasing amounts of interest. This essay approaches the epistemic foundations of ML-generated information specifically and medical knowledge generally to advocate differentiations of decision-making situations in clinical contexts regarding their necessary depth of insight into the process of information generation. Empirically accurate or reliable outcomes are sufficient for some decision situations in healthcare, whereas other clinical decisions require extensive insight into ML-generated outcomes because of their inherently normative implications.
Tài liệu tham khảo
Ahuja, A. S. (2019). The impact of artificial intelligence in medicine on the future role of the physician. PeerJ, 7, e7702. https://doi.org/10.7717/peerj.7702
Baggio, G., Corsini, A., Floreani, A., Giannini, S., & Zagonel, V. (2013). Gender medicine: A task for the third millennium. Clinical Chemistry and Laboratory Medicine, 51(4), 713–727. https://doi.org/10.1515/cclm-2012-0849
Bjerring, J. C., & Busch, J. (2021). Artificial intelligence and patient-centered decision-making. Philosophy & Technology, 34, 349–371. https://doi.org/10.1007/s13347-019-00391-6
Cabitza, F., Rasoini, R., & Gensini, G. F. (2017). Unintended consequences of machine learning in medicine. JAMA, 318(6), 517–518. https://doi.org/10.1001/jama.2017.7797
Cartwright, N. (2007a). Are RCTs the gold standard? BioSocieties, 2(2), 11–20. https://doi.org/10.1017/S1745855207005029
Cartwright, N. (2007b). Evidence-based policy: Where is our theory of evidence? Center for Philosophy of Natural and Social Science, London School of Economics, Technical Report 07/07.
Chakravartty, A. (2017). Scientific Realism. The Stanford Encyclopedia of Philosophy (Summer 2017 Edition), Retrieved January 6, 2022, from https://plato.stanford.edu/archives/sum2017/entries/scientific-realism/
De Fauw, J., Ledsam, J. R., Romera-Paredes, B., et al. (2018). Clinically applicable deep learning for diagnosis and referral in retinal disease. Nature Medicine, 24(9), 1342–1350. https://doi.org/10.1038/s41591-018-0107-6
de Regt, H. W., Leonelli, S., & Eigner, K. (Eds.). (2009). Scientific Understanding: Philosophical Perspectives. University of Pittsburgh Press.
Densen, P. (2011). Challenges and opportunities facing medical education. Transactions of the American Clinical and Climatological Association, 122, 48–58.
Durán, J. M., & Jongsma, K. R. (2021). Who is afraid of black box algorithms? On the epistemological and ethical basis of trust in medical AI. Journal of Medical Ethics, 47, 329–335. https://doi.org/10.1136/medethics-2020-106820
Esteva, A., Robicquet, A., Ramsundar, B., et al. (2019). A guide to deep learning in healthcare. Nature Medicine, 25, 24–29. https://doi.org/10.1038/s41591-018-0316-z
Floridi, L., Cowls, J., Beltrametti, M., Chatile, R., Chazerand, P., Dignum, V., Luetge, C., Madelin, R., Pagallo, U., Rossi, F., Schafer, B., Valcke, P., & Vayena, E. (2018). AI4People – An ethical framework for a good AI society: Opportunities, risks, principles, and recommendations. Minds & Machines, 28, 689–707. https://doi.org/10.1007/s11023-018-9482-5
Genin, K., Grote, T. (2021). Randomized controlled trials in medical AI. A methodological critique. Philosophiy of Medicine 2, 1–15. https://doi.org/10.5195/POM.2021.27.
Goldman, A. I. (2001). Experts: Which ones should you trust? Philosophy and Phenomenological Research, 63, 85–110.
Grimm, Stephen R. (2005). Understanding as an epistemic goal, Dissertation (University of Notre Dame).
Grimm, S. (2011). “Understanding”. In The Routledge Companion to Epistemology. Edited by S. Berneker D. Pritchard, 84–94. New York: Routledge, 2011.
Grimm, S. (Ed.). (2017). Making Sense of the World. Oxford University Press.
Grimm, S., Baumberger, C., & Ammon, S. (Eds.). (2017). Explaining understanding: New perspectives from epistemology and philosophy of science. Routledge.
Grote, T., & Berens, P. (2020). On the ethics of algorithmic decision-making in healthcare. Journal of Medical Ethics, 46, 205–211. https://doi.org/10.1136/medethics-2019-105586
Hardin, C. L., & Rosenberg, A. (1982). In Defence of Convergent Realism. Philosophy of Science, 49(4), 604–615. https://doi.org/10.1086/289080
Heinrichs, B., & Eickhoff, S. B. (2020). Your evidence? Machine learning algorithms for medical diagnosis and prediction. Human Brain Mapping, 41, 1435–1444. https://doi.org/10.1002/hbm.24886
Hinton, G. E. (2007). Learning multiple layers of representation. Trends in Cognitive Sciences, 11, 428–434. https://doi.org/10.1016/j.tics.2007.09.004
Holzinger, A., Carrington, A., & Müller, H. (2020). Measuring the quality of explanations: The system causability score (SCS). KI – Künstliche Intelligenz, 34, 193–198. https://doi.org/10.1007/s13218-020-00636-z.
Houssami, N., Lee, C. I., Buist, D. S. M., & Tao, D. (2017). Artificial intelligence for breast cancer screening: Opportunity or hype? The Breast, 36, 31–33.https://doi.org/10.1016/j.breast.2017.09.003.
Hutson, M. (2021). Lyin’ AIs: The opacity of artificial intelligence makes it hard to tell when decision-making is biased. IEEE Spectrum, 58(2), 40–45. https://doi.org/10.1109/MSPEC.2021.9340114
Johnson, K. W., Torres Soto, J., Glicksberg, B. S., Shameer, K., Miotto, R., Ali, M., Ashley, E., & Dudley, J. T. (2018). Artificial intelligence in cardiology. Journal of the American College of Cardiology, 71(23), 2668–2679. https://doi.org/10.1016/j.jacc.2018.03.521
Krishnan, M. (2020). Against interpretability: A Critical examination of the interpretability problem in machine learning. Philosophy & Technology, 33, 487–502. https://doi.org/10.1007/s13347-019-00372-9
Krittanawong, C., Zhang, H., Wang, Z., Aydar, M., & Kitai, T. (2017). Artificial intelligence in precision cardiovascular medicine. Journal of the American College of Cardiology, 69(21), 2657–2664. https://doi.org/10.1016/j.jacc.2017.03.571
Liu, X., Faes, L., Kale, A. U., Wagner, S. K., Fu, D. J., Bruynseels, A., Mahendiran, T., Moraes, G., Shamdas, M., Kern, C., Ledsam, J. R., Schmid, M. K., Balaskas, K., Topol, E. J., Bachmann, L. M., Keane, P. A., & Denniston, A. K. (2019). A comparison of deep learning performance against health-care professionals in detecting diseases from medical imaging: A systematic review and meta-analysis. The Lancet – Digital Health, 1(6), E271–E297. https://doi.org/10.1016/S2589-7500(19)30123-2.
London, A. J. (2019). Artificial intelligence and black-box. Medical decisions: Accuracy versus explainability. Hastings Center Report, 49(1), 15–21. https://doi.org/10.1002/hast.973.
Martini, C. (2020). The Epistemology of Expertise. In M. Fricker, P. J. Graham, D. Henderson, & N. J. L. L. Pedersen (Eds.), The Routledge Handbook of Social Epistemology (pp. 115–122). Routledge.
Molnar, C. (2021). Interpretable Machine Learning. A Guide for Making Black Box Models Explainable. Retrieved August 20, 2021, from https://christophm.github.io/interpretable-ml-book/.
Nagendran, M., Chen, Y., Lovejoy, C. A., Gordon, A. C., Komorowski, M., Harvey, H., Topol, E. J., Ioannidis, J. P. A., Collins, G. S., & Maruthappu, M. (2020). Artificial intelligence versus clinicians: Systematic review of design, reporting standards, and claims of deep learning studies. BMJ, 368, m689. https://doi.org/10.1136/bmj.m689
Orwat, C. (2019). Studie Diskriminierungsrisiken durch Verwendung von Algorithmen. Retrieved April 11, 2021, from: https://www.antidiskriminierungsstelle.de/SharedDocs/Downloads/DE/publikationen/Expertisen/Studie_Diskriminierungsrisiken_durch_Verwendung_von_Algorithmen.html.
Patel, S., Wang, J. V., Motaparthi, K., & Lee, J. B. (2021). Artificial intelligence in dermatology for the clinician. Clinics in Dermatology. In Press. https://doi.org/10.1016/j.clindermatol.2021.03.012.
Pearson, D. (2017). Artificial intelligence in radiology: the game-changer on everyone’s mind. Radiology business. Retrieved April 11, 2021, from: https://www.radiologybusiness.com/topics/technology-management/artificial-intelligence-radiology-game-changer-everyones-mind
Pritchard, D. (2009). Knowledge. Palgrave Macmillan.
Putnam, H. (1982). Three Kinds of Scientific Realism. Philosophical Quarterly, 32(128), 195–200. https://doi.org/10.2307/2219323
Rawlins, M. (2008). De testimonio: On the evidence for decisions about the use of therapeutic interventions. Lancet, 372(9656), 2152–2161. https://doi.org/10.1016/S0140-6736(08)61930-3
Robbins, S. (2019). A misdirected principle with a catch: Explicability for AI. Minds and Machines, 29, 495–514. https://doi.org/10.1007/s11023-019-09509-3
Rudin, C., & Radin, J. (2019). Why are we using black box models in AI when we don’t need to? A lesson from an explainable AI competition. Harvard Data Science Review, 1(2). https://doi.org/10.1162/99608f92.5a8a3a3d.
Salto-Tellez, M., Maxwell, P., & Hamilton, P. W. (2018). Artificial intelligence – The third revolution in pathology. Histopathology. https://doi.org/10.1111/his.13760
Schmidt-Erfurth, U., Sadeghipour, A., Gerendas, B. S., Waldstein, S. M., & Bogunović, H. (2018). Artificial intelligence in retina. Progress in Retinal and Eye Research, 67, 1–29. https://doi.org/10.1016/j.preteyeres.2018.07.004
Sedmak, C. (2003). Erkennen und Verstehen. Grundkurs Erkenntnistheorie und Hermeneutik. Tyrolia Innsbruck.
Seidlein, A. H., & Salloch, S. (2019). Illness and disease: An empirical-ethical viewpoint. BMC Medical Ethics, 20(1), 5. https://doi.org/10.1186/s12910-018-0341-y
Sim, I., Gorman, P., Greenes, R. A., Haynes, R. B., Kaplan, B., Lehmann, H., & Tang, P. C. (2001). Clinical Decision Support Systems for the Practice of Evidence-based Medicine. Journal of the American Medical Informatics Association, 8, 527–534. https://doi.org/ 10.1136/jamia.2001.0080527
Smith, P. (1998). Approximate truth and dynamical theories. British Journal for the Philosophy of Science, 49(2), 253–277. https://doi.org/10.1093/bjps/49.2.253
Smith, H. (2021). Clinical AI: Opacity, accountability, responsibility and liability. AI & Society. https://doi.org/10.1007/s00146-020-01019-6
Solomon, M. (2015). Making Medical Knowledge. Oxford University Press.
Spreckelsen, C., & Spitzer, K. (2008). Wissensbasen und Expertensysteme in der Medizin. KI-Ansätze zwischen klinischer Entscheidungsunterstützung und medizinischem Wissensmanagement. Medizinische Informatik. Vieweg + Teubner.
Tsamados, A., Aggarwal, N., Cowls, J., Morley, J., Roberts, H., Taddeo, M., & Floridi, L. (2021). The ethics of algorithms: Key problems and solutions. AI & Society. https://doi.org/10.1007/s00146-021-01154-8
Topol, E. J. (2019). High-performance medicine: The convergence of human and artificial intelligence. Nature Medicine, 25, 44–56. https://doi.org/10.1038/s41591-018-0300-7
Visani, G., Bagli, E., & Chesani, F. (2020). OptiLIME: Optimized LIME explanations for diagnostic computer algorithms. Proceedings of ACM Conference ’17. ACM New York.
Worrall, J. (2007). Evidence in medicine and evidence-based medicine. Philosophy Compass, 2(6), 981–1022. https://doi.org/10.1111/j.1747-9991.2007.00106.x
Zagzebski, L. (2009). On Epistemology. Wadsworth.
Zednik, C. (2021). Solving the black box problem: A normative framework for explainable artificial intelligence. Philosophy & Technology, 34, 265–288. https://doi.org/10.1007/s13347-019-00382-7
Zhou, X.-Y., Guo, Y., Shen, M., & Yang, G.-Z. (2020). Application of artificial intelligence in surgery. Frontiers in Medicine, 14, 417–430. https://doi.org/10.1007/s11684-020-0770-0.