Please use this identifier to cite or link to this item: http://dspace.iitrpr.ac.in:8080/xmlui/handle/123456789/4295
Full metadata record
DC FieldValueLanguage
dc.contributor.authorKamakshi, V.-
dc.contributor.authorKrishnan, N.C.-
dc.date.accessioned2022-12-09T08:02:25Z-
dc.date.available2022-12-09T08:02:25Z-
dc.date.issued2022-12-09-
dc.identifier.urihttp://localhost:8080/xmlui/handle/123456789/4295-
dc.description.abstractDomain adaptation techniques have contributed to the success of deep learning. Leveraging knowledge from an auxiliary source domain for learning in labeled data-scarce target domain is fundamental to domain adaptation. While these techniques result in increasing accuracy, the adaptation process, particularly the knowledge leveraged from the source domain, remains unclear. This paper proposes an explainable by design supervised domain adaptation framework - XSDA-Net. We integrate a case-based reasoning mechanism into the XSDA-Net to explain the prediction of a test instance in terms of similar-looking regions in the source and target train images. We empirically demonstrate the utility of the proposed framework by curating the domain adaptation settings on datasets popularly known to exhibit part-based explainability.en_US
dc.language.isoen_USen_US
dc.subjectExplainable by designen_US
dc.subjectInterpretable MLen_US
dc.subjectExplainable AIen_US
dc.subjectDomain adaptationen_US
dc.subjectExplainable domain adaptationen_US
dc.titleExplainable supervised domain adaptationen_US
dc.typeArticleen_US
Appears in Collections:Year-2022

Files in This Item:
File Description SizeFormat 
Full Text.pdf2.24 MBAdobe PDFView/Open    Request a copy


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.