ProtoVAE: A Trustworthy Self-Explainable Prototypical Variational Model
Permanent link
https://hdl.handle.net/10037/33160Date
2022-10-15Type
Journal articleTidsskriftartikkel
Peer reviewed
Author
Gautam, Srishti; Boubekki, Ahcene; Hansen, Stine; Salahuddin, Suaiba Amina; Jenssen, Robert; Hohne, Marina Marie-Claire; Kampffmeyer, MichaelAbstract
The need for interpretable models has fostered the development of self-explainable classifiers. Prior approaches are either based on multi-stage optimization schemes, impacting the predictive performance of the model, or produce explanations that are not transparent, trustworthy or do not capture the diversity of the data. To address these shortcomings, we propose ProtoVAE, a variational autoencoder-based framework that learns class-specific prototypes in an end-to-end manner and enforces trustworthiness and diversity by regularizing the representation space and introducing an orthonormality constraint. Finally, the model is designed to be transparent by directly incorporating the prototypes into the decision process. Extensive comparisons with previous self-explainable approaches demonstrate the superiority of ProtoVAE, highlighting its ability to generate trustworthy and diverse explanations, while not degrading predictive performance.
Description
Source at https://nips.cc/.
Citation
Gautam S, Boubekki A, Hansen S, Salahuddin SA, Jenssen R, Hohne MM, Kampffmeyer MC. ProtoVAE: A Trustworthy Self-Explainable Prototypical Variational Model. Advances in Neural Information Processing Systems. 2022Metadata
Show full item recordCollections
Copyright 2022 The Author(s)