INSTITUTIONAL DIGITAL REPOSITORY

PACE: posthoc Architecture-Agnostic concept extractor for explaining CNNs

Show simple item record

dc.contributor.author Kamakshi, V.
dc.contributor.author Gupta, U.
dc.contributor.author Krishnan, N. C.
dc.date.accessioned 2021-11-22T09:36:40Z
dc.date.available 2021-11-22T09:36:40Z
dc.date.issued 2021-11-22
dc.identifier.uri http://localhost:8080/xmlui/handle/123456789/3226
dc.description.abstract Deep CNNs, though have achieved the state of the art performance in image classification tasks, remain a black-box to a human using them. There is a growing interest in explaining the working of these deep models to improve their trustworthiness. In this paper, we introduce a Posthoc Architecture-agnostic Concept Extractor (PACE) that automatically extracts smaller sub-regions of the image called concepts relevant to the black-box prediction. PACE tightly integrates the faithfulness of the explanatory framework to the black-box model. To the best of our knowledge, this is the first work that extracts class-specific discriminative concepts in a posthoc manner automatically. The PACE framework is used to generate explanations for two different CNN architectures trained for classifying the AWA2 and Imagenet-Birds datasets. Extensive human subject experiments are conducted to validate the human interpretability and consistency of the explanations extracted by PACE. The results from these experiments suggest that over 72% of the concepts extracted by PACE are human interpretable. en_US
dc.language.iso en_US en_US
dc.subject XAI en_US
dc.subject posthoc explanations en_US
dc.subject concept-based explanations en_US
dc.subject image classifier explanations en_US
dc.title PACE: posthoc Architecture-Agnostic concept extractor for explaining CNNs en_US
dc.type Article en_US


Files in this item

This item appears in the following Collection(s)

Show simple item record

Search DSpace


Advanced Search

Browse

My Account