Title data
Kästner, Lena ; Crook, Barnaby:
Explaining AI through mechanistic interpretability.
In: European Journal for Philosophy of Science.
Vol. 14
(2024)
Issue 4
.
- 52.
ISSN 1879-4920
DOI: https://doi.org/10.1007/s13194-024-00614-4
Project information
Project financing: |
VolkswagenStiftung |
---|
Abstract in another language
Recent work in explainable artificial intelligence (XAI) attempts to render opaque AI systems understandable through a divide-and-conquer strategy. However, this fails to illuminate how trained AI systems work as a whole. Precisely this kind of functional understanding is needed, though, to satisfy important societal desiderata such as safety. To remedy this situation, we argue, AI researchers should seek mechanistic interpretability, viz. apply coordinated discovery strategies familiar from the life sciences to uncover the functional organisation of complex AI systems. Additionally, theorists should accommodate for the unique costs and benefits of such strategies in their portrayals of XAI research.
Further data
Item Type: | Article in a journal |
---|---|
Refereed: | Yes |
Keywords: | AI; ANN; Deep learning; Discovery; Explanation; Mechanistic; interpretability; XAI |
Institutions of the University: | Faculties > Faculty of Cultural Studies > Department of Philosophy > Chair Philosophy, Computer Science and Artificial Intelligence > Chair Philosophy, Computer Science and Artificial Intelligence - Univ.-Prof. Dr. Lena Kästner Faculties Faculties > Faculty of Cultural Studies Faculties > Faculty of Cultural Studies > Department of Philosophy Faculties > Faculty of Cultural Studies > Department of Philosophy > Chair Philosophy, Computer Science and Artificial Intelligence |
Result of work at the UBT: | Yes |
DDC Subjects: | 100 Philosophy and psychology > 100 Philosophy |
Date Deposited: | 08 Mar 2025 22:00 |
Last Modified: | 10 Mar 2025 07:06 |
URI: | https://eref.uni-bayreuth.de/id/eprint/92739 |