File Download

There are no files associated with this item.

  Links for fulltext
     (May Require Subscription)
Supplementary

Article: Gradient-Based Instance-Specific Visual Explanations for Object Specification and Object Discrimination

TitleGradient-Based Instance-Specific Visual Explanations for Object Specification and Object Discrimination
Authors
KeywordsDeep learning
explainable AI
explaining object detection
gradient-based explanation
human eye gaze
instance-level explanation
knowledge distillation
non-maximum suppression
object discrimination
object specification
Issue Date22-Mar-2024
PublisherInstitute of Electrical and Electronics Engineers
Citation
IEEE Transactions on Pattern Analysis and Machine Intelligence, 2024, v. 46, n. 9, p. 5967-5985 How to Cite?
Abstract

We propose the gradient-weighted Object Detector Activation Maps (ODAM), a visual explanation technique for interpreting the predictions of object detectors. Utilizing the gradients of detector targets flowing into the intermediate feature maps, ODAM produces heat maps that show the influence of regions on the detector's decision for each predicted attribute. Compared to previous works on classification activation maps (CAM), ODAM generates instance-specific explanations rather than class-specific ones. We show that ODAM is applicable to one-stage, two-stage, and transformer-based detectors with different types of detector backbones and heads, and produces higher-quality visual explanations than the state-of-the-art in terms of both effectiveness and efficiency. We discuss two explanation tasks for object detection: 1) object specification: what is the important region for the prediction? 2) object discrimination: which object is detected? Aiming at these two aspects, we present a detailed analysis of the visual explanations of detectors and carry out extensive experiments to validate the effectiveness of the proposed ODAM. Furthermore, we investigate user trust on the explanation maps, how well the visual explanations of object detectors agrees with human explanations, as measured through human eye gaze, and whether this agreement is related with user trust. Finally, we also propose two applications, ODAM-KD and ODAM-NMS, based on these two abilities of ODAM. ODAM-KD utilizes the object specification of ODAM to generate top-down attention for key predictions and instruct the knowledge distillation of object detection. ODAM-NMS considers the location of the model's explanation for each prediction to distinguish the duplicate detected objects. A training scheme, ODAM-Train, is proposed to improve the quality on object discrimination, and help with ODAM-NMS.


Persistent Identifierhttp://hdl.handle.net/10722/351192
ISSN
2023 Impact Factor: 20.8
2023 SCImago Journal Rankings: 6.158

 

DC FieldValueLanguage
dc.contributor.authorZhao, Chenyang-
dc.contributor.authorHsiao, Janet H-
dc.contributor.authorChan, Antoni B-
dc.date.accessioned2024-11-13T00:36:06Z-
dc.date.available2024-11-13T00:36:06Z-
dc.date.issued2024-03-22-
dc.identifier.citationIEEE Transactions on Pattern Analysis and Machine Intelligence, 2024, v. 46, n. 9, p. 5967-5985-
dc.identifier.issn0162-8828-
dc.identifier.urihttp://hdl.handle.net/10722/351192-
dc.description.abstract<p>We propose the gradient-weighted Object Detector Activation Maps (ODAM), a visual explanation technique for interpreting the predictions of object detectors. Utilizing the gradients of detector targets flowing into the intermediate feature maps, ODAM produces heat maps that show the influence of regions on the detector's decision for each predicted attribute. Compared to previous works on classification activation maps (CAM), ODAM generates instance-specific explanations rather than class-specific ones. We show that ODAM is applicable to one-stage, two-stage, and transformer-based detectors with different types of detector backbones and heads, and produces higher-quality visual explanations than the state-of-the-art in terms of both effectiveness and efficiency. We discuss two explanation tasks for object detection: 1) object specification: what is the important region for the prediction? 2) object discrimination: which object is detected? Aiming at these two aspects, we present a detailed analysis of the visual explanations of detectors and carry out extensive experiments to validate the effectiveness of the proposed ODAM. Furthermore, we investigate user trust on the explanation maps, how well the visual explanations of object detectors agrees with human explanations, as measured through human eye gaze, and whether this agreement is related with user trust. Finally, we also propose two applications, ODAM-KD and ODAM-NMS, based on these two abilities of ODAM. ODAM-KD utilizes the object specification of ODAM to generate top-down attention for key predictions and instruct the knowledge distillation of object detection. ODAM-NMS considers the location of the model's explanation for each prediction to distinguish the duplicate detected objects. A training scheme, ODAM-Train, is proposed to improve the quality on object discrimination, and help with ODAM-NMS.</p>-
dc.languageeng-
dc.publisherInstitute of Electrical and Electronics Engineers-
dc.relation.ispartofIEEE Transactions on Pattern Analysis and Machine Intelligence-
dc.rightsThis work is licensed under a Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 International License.-
dc.subjectDeep learning-
dc.subjectexplainable AI-
dc.subjectexplaining object detection-
dc.subjectgradient-based explanation-
dc.subjecthuman eye gaze-
dc.subjectinstance-level explanation-
dc.subjectknowledge distillation-
dc.subjectnon-maximum suppression-
dc.subjectobject discrimination-
dc.subjectobject specification-
dc.titleGradient-Based Instance-Specific Visual Explanations for Object Specification and Object Discrimination-
dc.typeArticle-
dc.identifier.doi10.1109/TPAMI.2024.3380604-
dc.identifier.scopuseid_2-s2.0-85188896489-
dc.identifier.volume46-
dc.identifier.issue9-
dc.identifier.spage5967-
dc.identifier.epage5985-
dc.identifier.eissn1939-3539-
dc.identifier.issnl0162-8828-

Export via OAI-PMH Interface in XML Formats


OR


Export to Other Non-XML Formats