Component based comparative analysis of each module in image captioning
Image captioning is a task to generate a new caption using the training data of the image and caption. Since existing deep learning is a black-box model, it is crucial to analyze the influence on each module for understanding the model. In this paper, we analyze the impact of the five modules and do...
Main Authors: | Seoung-Ho Choi, Seoung Yeon Jo, Sung Hoon Jung |
---|---|
Format: | Article |
Language: | English |
Published: |
Elsevier
2021-03-01
|
Series: | ICT Express |
Subjects: | |
Online Access: | http://www.sciencedirect.com/science/article/pii/S2405959520301429 |
Similar Items
-
Text Augmentation Using BERT for Image Captioning
by: Viktar Atliha, et al.
Published: (2020-08-01) -
#PraCegoVer: A Large Dataset for Image Captioning in Portuguese
by: Gabriel Oliveira dos Santos, et al.
Published: (2022-01-01) -
Full-Memory Transformer for Image Captioning
by: Tongwei Lu, et al.
Published: (2023-01-01) -
Generalized Image Captioning for Multilingual Support
by: Suhyun Cho, et al.
Published: (2023-02-01) -
VAA: Visual Aligning Attention Model for Remote Sensing Image Captioning
by: Zhengyuan Zhang, et al.
Published: (2019-01-01)