Explainable Models with Consistent Interpretations

dc.contributor.authorPillai, Vipin
dc.contributor.authorPirsiavash, Hamed
dc.date.accessioned2021-02-19T16:44:14Z
dc.date.available2021-02-19T16:44:14Z
dc.date.issued2021
dc.description.abstractGiven the widespread deployment of black box deep neural networks in computer vision applications, the interpretability aspect of these black box systems has recently gained traction. Various methods have been proposed to explain the results of such deep neural networks. However, some recent works have shown that such explanation methods are biased and do not produce consistent interpretations. Hence, rather than introducing a novel explanation method, we learn models that are encouraged to be interpretable given an explanation method. We use Grad-CAM as the explanation algorithm and encourage the network to learn consistent interpretations along with maximizing the log-likelihood of the correct class. We show that our method outperforms the baseline on the pointing game evaluation on ImageNet and MS-COCO datasets respectively. We also introduce new evaluation metrics that penalize the saliency map if it lies outside the ground truth bounding box or segmentation mask, and show that our method outperforms the baseline on these metrics as well. Moreover, our model trained with interpretation consistency generalizes to other explanation algorithms on all the evaluation metrics.en_US
dc.description.sponsorshipThis material is based upon work partially supported by the United States Air Force under Contract No. FA8750-19-C-0098, funding from NSF grant number 1845216, SAP SE, and Northrop Grumman. Any opinions, findings, and conclusions or recommendations expressed in this material are those of the authors and do not necessarily reflect the views of the United States Air Force, DARPA, or other funding agencies.en_US
dc.description.urihttps://www.aaai.org/AAAI21Papers/AAAI-8236.PillaiV.pdfen_US
dc.format.extent9 pagesen_US
dc.genrejournal articlesen_US
dc.genrepreprints
dc.identifierdoi:10.13016/m2orpw-45bj
dc.identifier.citationPillai, Vipin; Pirsiavash, Hamed; Explainable Models with Consistent Interpretations (2021); https://www.aaai.org/AAAI21Papers/AAAI-8236.PillaiV.pdfen_US
dc.identifier.urihttp://hdl.handle.net/11603/21053
dc.identifier.urihttps://doi.org/10.1609/aaai.v35i3.16344
dc.language.isoen_USen_US
dc.publisherAAAI
dc.relation.isAvailableAtThe University of Maryland, Baltimore County (UMBC)
dc.relation.ispartofUMBC Computer Science and Electrical Engineering Department Collection
dc.relation.ispartofUMBC Faculty Collection
dc.relation.ispartofUMBC Student Collection
dc.rightsThis item is likely protected under Title 17 of the U.S. Copyright Law. Unless on a Creative Commons license, for uses protected by Copyright Law, contact the copyright holder or the author.
dc.subjectdeep neural networksen_US
dc.subjectevaluation metricsen_US
dc.subjectalgorithmsen_US
dc.titleExplainable Models with Consistent Interpretationsen_US
dc.typeTexten_US

Files

Original bundle
Now showing 1 - 1 of 1
Loading...
Thumbnail Image
Name:
AAAI-8236.PillaiV.pdf
Size:
2.2 MB
Format:
Adobe Portable Document Format
Description:
License bundle
Now showing 1 - 1 of 1
No Thumbnail Available
Name:
license.txt
Size:
2.56 KB
Format:
Item-specific license agreed upon to submission
Description: