Please use this identifier to cite or link to this item:
http://dspace.iitrpr.ac.in:8080/xmlui/handle/123456789/4295
Full metadata record
DC Field | Value | Language |
---|---|---|
dc.contributor.author | Kamakshi, V. | - |
dc.contributor.author | Krishnan, N.C. | - |
dc.date.accessioned | 2022-12-09T08:02:25Z | - |
dc.date.available | 2022-12-09T08:02:25Z | - |
dc.date.issued | 2022-12-09 | - |
dc.identifier.uri | http://localhost:8080/xmlui/handle/123456789/4295 | - |
dc.description.abstract | Domain adaptation techniques have contributed to the success of deep learning. Leveraging knowledge from an auxiliary source domain for learning in labeled data-scarce target domain is fundamental to domain adaptation. While these techniques result in increasing accuracy, the adaptation process, particularly the knowledge leveraged from the source domain, remains unclear. This paper proposes an explainable by design supervised domain adaptation framework - XSDA-Net. We integrate a case-based reasoning mechanism into the XSDA-Net to explain the prediction of a test instance in terms of similar-looking regions in the source and target train images. We empirically demonstrate the utility of the proposed framework by curating the domain adaptation settings on datasets popularly known to exhibit part-based explainability. | en_US |
dc.language.iso | en_US | en_US |
dc.subject | Explainable by design | en_US |
dc.subject | Interpretable ML | en_US |
dc.subject | Explainable AI | en_US |
dc.subject | Domain adaptation | en_US |
dc.subject | Explainable domain adaptation | en_US |
dc.title | Explainable supervised domain adaptation | en_US |
dc.type | Article | en_US |
Appears in Collections: | Year-2022 |
Files in This Item:
File | Description | Size | Format | |
---|---|---|---|---|
Full Text.pdf | 2.24 MB | Adobe PDF | View/Open Request a copy |
Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.