Explainable Machine Learning and its Limitations

DSpace Repositorium (Manakin basiert)

Zur Kurzanzeige

dc.contributor.advisor von Luxburg, Ulrike (Prof. Dr.)
dc.contributor.author Bordt, Sebastian
dc.date.accessioned 2023-11-14T10:42:46Z
dc.date.available 2023-11-14T10:42:46Z
dc.date.issued 2023-11-14
dc.identifier.uri http://hdl.handle.net/10900/147741
dc.identifier.uri http://nbn-resolving.de/urn:nbn:de:bsz:21-dspace-1477414 de_DE
dc.identifier.uri http://dx.doi.org/10.15496/publikation-89082
dc.description.abstract In the last decade, machine learning evolved from a sub-field of computer science into one of the most impactful scientific disciplines of our time. While this has brought impressive scientific advances, there are now increasing concerns about the applications of artificial intelligence systems in societal contexts. Many concerns are rooted in the fact that machine learning models can be incredibly opaque. To overcome this problem, the nascent field of explainable machine learning attempts to provide human-understandable explanations for the behavior of complex models. After an initial period of method development and excitement, researchers in this field have now recognized the many difficulties inherent in faithfully explaining complex models. In this thesis, we review the developments within the first decade of explainable machine learning. We outline the main motivations for explainable machine learning, as well as some of the debates within the field. We also make three specific contributions that attempt to clarify what is and is not possible when explaining complex models. The first part of the thesis studies the learning dynamics of the human-machine decision making problem. We show how this learning problem is different from other forms of collaborative decision making, and derive conditions under which it can be efficiently solved. We also clarify the role of algorithmic explanations in this setup. In the second part of the thesis, we study the suitability of local post-hoc explanation algorithms in societal contexts. Focusing on the draft EU Artificial Intelligence Act, we argue that these methods are unable to fulfill the transparency objectives that are inherent in the law. Our results also suggest that regulating artificial intelligence systems implicitly via their explanations is unlikely to succeed with currently available methods. In the third part of the thesis, we provide a detailed mathematical analysis of Shapley Values, a prominent model explanation technique, and show how it is connected with Generalized Additive Models, a popular class of interpretable models. The last part of the thesis serves as an interesting case study of a connection between a post-hoc method and a class of interpretable models. en
dc.language.iso en de_DE
dc.publisher Universität Tübingen de_DE
dc.rights ubt-podok de_DE
dc.rights.uri http://tobias-lib.uni-tuebingen.de/doku/lic_mit_pod.php?la=de de_DE
dc.rights.uri http://tobias-lib.uni-tuebingen.de/doku/lic_mit_pod.php?la=en en
dc.subject.classification Informatik de_DE
dc.subject.ddc 004 de_DE
dc.subject.other Machine Learning de_DE
dc.subject.other Machine Learning en
dc.title Explainable Machine Learning and its Limitations en
dc.type PhDThesis de_DE
dcterms.dateAccepted 2023-09-27
utue.publikation.fachbereich Informatik de_DE
utue.publikation.fakultaet 7 Mathematisch-Naturwissenschaftliche Fakultät de_DE
utue.publikation.noppn yes de_DE

Dateien:

Das Dokument erscheint in:

Zur Kurzanzeige