dc.contributor.author |
Kamakshi, V. |
|
dc.contributor.author |
Gupta, U. |
|
dc.contributor.author |
Krishnan, N. C. |
|
dc.date.accessioned |
2021-11-22T09:36:40Z |
|
dc.date.available |
2021-11-22T09:36:40Z |
|
dc.date.issued |
2021-11-22 |
|
dc.identifier.uri |
http://localhost:8080/xmlui/handle/123456789/3226 |
|
dc.description.abstract |
Deep CNNs, though have achieved the state of
the art performance in image classification tasks, remain a
black-box to a human using them. There is a growing interest
in explaining the working of these deep models to improve
their trustworthiness. In this paper, we introduce a Posthoc
Architecture-agnostic Concept Extractor (PACE) that automatically extracts smaller sub-regions of the image called concepts
relevant to the black-box prediction. PACE tightly integrates
the faithfulness of the explanatory framework to the black-box
model. To the best of our knowledge, this is the first work
that extracts class-specific discriminative concepts in a posthoc
manner automatically. The PACE framework is used to generate
explanations for two different CNN architectures trained for
classifying the AWA2 and Imagenet-Birds datasets. Extensive
human subject experiments are conducted to validate the human
interpretability and consistency of the explanations extracted by
PACE. The results from these experiments suggest that over 72%
of the concepts extracted by PACE are human interpretable. |
en_US |
dc.language.iso |
en_US |
en_US |
dc.subject |
XAI |
en_US |
dc.subject |
posthoc explanations |
en_US |
dc.subject |
concept-based explanations |
en_US |
dc.subject |
image classifier explanations |
en_US |
dc.title |
PACE: posthoc Architecture-Agnostic concept extractor for explaining CNNs |
en_US |
dc.type |
Article |
en_US |