Towards Human-Like Interpretable Object Detection Via Spatial Relation Encoding

Cited 5 time in webofscience Cited 4 time in scopus
  • Hit : 1463
  • Download : 0
The performance of recent deep neural networks in various computer vision areas such as object detection has increased significantly. Along with such advances, attempts to visualize and interpret the networks have been made in order to understand how a network predicts a certain result. However, there is a lack of research on ways to improve the interpretability of networks' features. In this paper, we propose a spatial relation reasoning (SRR) framework to encode interpretable networks' features, especially an object detector, by mimicking the human visual cognition system. The SRR consists of the spatial feature encoder (SFE) and the graph-based spatial relation encoder (GSRE) to consider spatial relationships between different parts of an object. So that, object detectors can encode spatially-related object features enabling humanlike visual interpretation. We verified the proposed framework with general object detectors on public datasets-PAS-CAL VOC and MS COCO.
Publisher
IEEE Signal Processing Society
Issue Date
2020-10-25
Language
English
Citation

IEEE International Conference on Image Processing (ICIP) 2020, pp.3284 - 3288

ISSN
1522-4880
DOI
10.1109/ICIP40778.2020.9190724
URI
http://hdl.handle.net/10203/274235
Appears in Collection
EE-Conference Papers(학술회의논문)
Files in This Item
There are no files associated with this item.
This item is cited by other documents in WoS
⊙ Detail Information in WoSⓡ Click to see webofscience_button
⊙ Cited 5 items in WoS Click to see citing articles in records_button

qr_code

  • mendeley

    citeulike


rss_1.0 rss_2.0 atom_1.0