Please use this identifier to cite or link to this item:
http://dspace.iitrpr.ac.in:8080/xmlui/handle/123456789/3944
Full metadata record
DC Field | Value | Language |
---|---|---|
dc.contributor.author | Sharma, R. | - |
dc.contributor.author | Reddy, N. | - |
dc.contributor.author | Kamakshi, V. | - |
dc.contributor.author | Krishnan, N.C. | - |
dc.contributor.author | Jain, S. | - |
dc.date.accessioned | 2022-09-03T08:47:42Z | - |
dc.date.available | 2022-09-03T08:47:42Z | - |
dc.date.issued | 2022-09-03 | - |
dc.identifier.uri | http://localhost:8080/xmlui/handle/123456789/3944 | - |
dc.description.abstract | The paper introduces a novel framework for extracting model-agnostic human interpretable rules to explain a classifier’s output. The human interpretable rule is defined as an axis-aligned hyper-cuboid containing the instance for which the classification decision has to be explained. The proposed procedure finds the largest (high coverage) axis-aligned hyper-cuboid such that a high percentage of the instances in the hyper-cuboid have the same class label as the instance being explained (high precision). Novel approximations to the coverage and precision measures in terms of the parameters of the hyper-cuboid are defined. They are maximized using gradient-based optimizers. The quality of the approximations is rigorously analyzed theoretically and experimentally. Heuristics for simplifying the generated explanations for achieving better interpretability and a greedy selection algorithm that combines the local explanations for creating global explanations for the model covering a large part of the instance space are also proposed. The framework is model agnostic, can be applied to any arbitrary classifier, and all types of attributes (including continuous, ordered, and unordered discrete). The wide-scale applicability of the framework is validated on a variety of synthetic and real-world datasets from different domains (tabular, text, and image). | en_US |
dc.language.iso | en_US | en_US |
dc.subject | Explainable models | en_US |
dc.subject | Interpretable machine learning | en_US |
dc.subject | Rule based explanations | en_US |
dc.title | MAIRE - A model-agnostic interpretable rule extraction procedure for explaining classifiers | en_US |
dc.type | Article | en_US |
Appears in Collections: | Year-2021 |
Files in This Item:
File | Description | Size | Format | |
---|---|---|---|---|
Full Text.pdf | 1.76 MB | Adobe PDF | View/Open Request a copy |
Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.