Explainable Image Caption Generator Using Attention and Bayesian Inference

Cited 0 time in webofscience Cited 4 time in scopus
  • Hit : 150
  • Download : 0
Image captioning is the task of generating textual descriptions of a given image, requiring techniques of computer vision and natural language processing. Recent models have utilized deep learning techniques to this task to gain performance improvement. However, these models can neither distinguish more important objects than others in a given image, nor explain the reasons why specific words have been selected when generating captions. To overcome these limitations, this paper proposes an explainable image captioning model, which generates a caption by indicating specific objects in a given image and providing visual explanation using them. The model has been evaluated with datasets such as MSCOCO, Flickr8K, and Flickr30K, and some qualitative results are presented to show the effectiveness of the proposed model.
Publisher
Institute of Electrical and Electronics Engineers Inc.
Issue Date
2018-12
Language
English
Citation

2018 International Conference on Computational Science and Computational Intelligence, CSCI 2018, pp.478 - 481

DOI
10.1109/csci46756.2018.00098
URI
http://hdl.handle.net/10203/277247
Appears in Collection
CS-Conference Papers(학술회의논문)
Files in This Item
There are no files associated with this item.

qr_code

  • mendeley

    citeulike


rss_1.0 rss_2.0 atom_1.0