From Explanations to Interpretability and Back

Räz, Tim (2024). From Explanations to Interpretability and Back (Submitted). In: Philosophy of Science for Machine Learning: Core Issues, New Perspectives. Synthese Library. Springer

[img] Text
raez_ml_expl_philsci.pdf - Submitted Version
Restricted to registered users only
Available under License Publisher holds Copyright.

Download (329kB) | Request a copy

This chapter discusses connections between interpretability of machine learning and (scientific and mathematical) explanations, provides novel perspectives on interpretability, and highlights under-explored issues. Interpretability types are proposed: kinds of interpretability should be distinguished using both the parts of ML we want to explain and the parts of ML we use to explain. It is argued that not all explanations are contrastive, and that we should also consider contrasts with respect to models and data, not only with respect to inputs. Theoretical explanations are highlighted; they include issues like generalization, optimization, and expressivity. It is proposed that there are two threats to the objectivity of explanations: One from radical subject-dependence, the other from a lack of factivity. Finally, pluralism is advocated: There are different notions of interpretability and different notions of (scientific and mathematical) explanations. However, the heterogeneity of one area does not transfer to the other in a straightforward manner.

Item Type:

Book Section (Book Chapter)

Division/Institute:

06 Faculty of Humanities > Department of Art and Cultural Studies > Institute of Philosophy

UniBE Contributor:

Räz, Tim

Subjects:

100 Philosophy
600 Technology > 620 Engineering

Series:

Synthese Library

Publisher:

Springer

Language:

English

Submitter:

Tim Räz

Date Deposited:

08 Mar 2024 15:05

Last Modified:

08 Mar 2024 15:05

Related URLs:

BORIS DOI:

10.48350/193896

URI:

https://boris.unibe.ch/id/eprint/193896

Actions (login required)

Edit item Edit item
Provide Feedback