Vis enkel innførsel

dc.contributor.authorGautam, Srishti
dc.contributor.authorHohne, Marina Marie-Claire
dc.contributor.authorHansen, Stine
dc.contributor.authorJenssen, Robert
dc.contributor.authorKampffmeyer, Michael
dc.date.accessioned2022-11-30T09:32:24Z
dc.date.available2022-11-30T09:32:24Z
dc.date.issued2022-11-12
dc.description.abstractCurrent machine learning models have shown high efficiency in solving a wide variety of real-world problems. However, their black box character poses a major challenge for the comprehensibility and traceability of the underlying decision-making strategies. As a remedy, numerous post-hoc and self-explanation methods have been developed to interpret the models’ behavior. Those methods, in addition, enable the identification of artifacts that, inherent in the training data, can be erroneously learned by the model as class-relevant features. In this work, we provide a detailed case study of a representative for the state-of-the-art self-explaining network, ProtoPNet, in the presence of a spectrum of artifacts. Accordingly, we identify the main drawbacks of ProtoPNet, especially its coarse and spatially imprecise explanations. We address these limitations by introducing Prototypical Relevance Propagation (PRP), a novel method for generating more precise model-aware explanations. Furthermore, in order to obtain a clean, artifact-free dataset, we propose to use multi-view clustering strategies for segregating the artifact images using the PRP explanations, thereby suppressing the potential artifact learning in the models.en_US
dc.identifier.citationGautam S, Hohne MM, Hansen S, Jenssen R, Kampffmeyer MC. This looks more like that: Enhancing Self-Explaining Models by Prototypical Relevance Propagation. Pattern Recognition. 2022en_US
dc.identifier.cristinIDFRIDAID 2084854
dc.identifier.doihttps://doi.org/10.1016/j.patcog.2022.109172
dc.identifier.issn0031-3203
dc.identifier.issn1873-5142
dc.identifier.urihttps://hdl.handle.net/10037/27611
dc.language.isoengen_US
dc.publisherElsevieren_US
dc.relation.ispartofGautam, S. (2024). Towards Interpretable, Trustworthy and Reliable AI. (Doctoral thesis). <a href=https://hdl.handle.net/10037/33143>https://hdl.handle.net/10037/33143</a>.
dc.relation.journalPattern Recognition
dc.rights.accessRightsopenAccessen_US
dc.rights.holderCopyright 2022 The Author(s)en_US
dc.rights.urihttps://creativecommons.org/licenses/by/4.0en_US
dc.rightsAttribution 4.0 International (CC BY 4.0)en_US
dc.titleThis looks more like that: Enhancing Self-Explaining Models by Prototypical Relevance Propagationen_US
dc.type.versionpublishedVersionen_US
dc.typeJournal articleen_US
dc.typeTidsskriftartikkelen_US
dc.typePeer revieweden_US


Tilhørende fil(er)

Thumbnail

Denne innførselen finnes i følgende samling(er)

Vis enkel innførsel

Attribution 4.0 International (CC BY 4.0)
Med mindre det står noe annet, er denne innførselens lisens beskrevet som Attribution 4.0 International (CC BY 4.0)