Controllable Image Captioning with Feature Refinement and Multilayer Fusion
Image captioning is the task of automatically generating a description of an image. Traditional image captioning models tend to generate a sentence describing the most conspicuous objects, but fail to describe a desired region or object as human. In order to generate sentences based on a given targe...
Main Authors: | , , , , , , , |
---|---|
Format: | Article |
Language: | English |
Published: |
MDPI AG
2023-04-01
|
Series: | Applied Sciences |
Subjects: | |
Online Access: | https://www.mdpi.com/2076-3417/13/8/5020 |
_version_ | 1797606491914502144 |
---|---|
author | Sen Du Hong Zhu Yujia Zhang Dong Wang Jing Shi Nan Xing Guangfeng Lin Huiyu Zhou |
author_facet | Sen Du Hong Zhu Yujia Zhang Dong Wang Jing Shi Nan Xing Guangfeng Lin Huiyu Zhou |
author_sort | Sen Du |
collection | DOAJ |
description | Image captioning is the task of automatically generating a description of an image. Traditional image captioning models tend to generate a sentence describing the most conspicuous objects, but fail to describe a desired region or object as human. In order to generate sentences based on a given target, understanding the relationships between particular objects and describing them accurately is central to this task. In detail, information-augmented embedding is used to add prior information to each object, and a new Multi-Relational Weighted Graph Convolutional Network (MR-WGCN) is designed for fusing the information of adjacent objects. Then, a dynamic attention decoder module selectively focuses on particular objects or semantic contents. Finally, the model is optimized by similarity loss. The experiment on MSCOCO Entities demonstrates that IANR obtains, to date, the best published CIDEr performance of 124.52% on the Karpathy test split. Extensive experiments and ablations on both the MSCOCO Entities and the Flickr30k Entities demonstrate the effectiveness of each module. Meanwhile, IANR achieves better accuracy and controllability than the state-of-the-art models under the widely used evaluation metric. |
first_indexed | 2024-03-11T05:15:57Z |
format | Article |
id | doaj.art-5f3b030084bd4a23a45a90d5634c2466 |
institution | Directory Open Access Journal |
issn | 2076-3417 |
language | English |
last_indexed | 2024-03-11T05:15:57Z |
publishDate | 2023-04-01 |
publisher | MDPI AG |
record_format | Article |
series | Applied Sciences |
spelling | doaj.art-5f3b030084bd4a23a45a90d5634c24662023-11-17T18:12:32ZengMDPI AGApplied Sciences2076-34172023-04-01138502010.3390/app13085020Controllable Image Captioning with Feature Refinement and Multilayer FusionSen Du0Hong Zhu1Yujia Zhang2Dong Wang3Jing Shi4Nan Xing5Guangfeng Lin6Huiyu Zhou7School of Automation and Information Engineering, Xi’an University of Technology, Xi’an 710048, ChinaSchool of Automation and Information Engineering, Xi’an University of Technology, Xi’an 710048, ChinaSchool of Automation and Information Engineering, Xi’an University of Technology, Xi’an 710048, ChinaSchool of Automation and Information Engineering, Xi’an University of Technology, Xi’an 710048, ChinaSchool of Automation and Information Engineering, Xi’an University of Technology, Xi’an 710048, ChinaSchool of Automation and Information Engineering, Xi’an University of Technology, Xi’an 710048, ChinaSchool of Printing, Packaging and Digital Media, Xi’an University of Technology, Xi’an 710054, ChinaSchool of Computing and Mathematical Sciences, University of Leicester, University Road, Leicester LE1 7RH, UKImage captioning is the task of automatically generating a description of an image. Traditional image captioning models tend to generate a sentence describing the most conspicuous objects, but fail to describe a desired region or object as human. In order to generate sentences based on a given target, understanding the relationships between particular objects and describing them accurately is central to this task. In detail, information-augmented embedding is used to add prior information to each object, and a new Multi-Relational Weighted Graph Convolutional Network (MR-WGCN) is designed for fusing the information of adjacent objects. Then, a dynamic attention decoder module selectively focuses on particular objects or semantic contents. Finally, the model is optimized by similarity loss. The experiment on MSCOCO Entities demonstrates that IANR obtains, to date, the best published CIDEr performance of 124.52% on the Karpathy test split. Extensive experiments and ablations on both the MSCOCO Entities and the Flickr30k Entities demonstrate the effectiveness of each module. Meanwhile, IANR achieves better accuracy and controllability than the state-of-the-art models under the widely used evaluation metric.https://www.mdpi.com/2076-3417/13/8/5020controllable image captioninginformation-augmented embeddingMR-WGCNsimilarity loss |
spellingShingle | Sen Du Hong Zhu Yujia Zhang Dong Wang Jing Shi Nan Xing Guangfeng Lin Huiyu Zhou Controllable Image Captioning with Feature Refinement and Multilayer Fusion Applied Sciences controllable image captioning information-augmented embedding MR-WGCN similarity loss |
title | Controllable Image Captioning with Feature Refinement and Multilayer Fusion |
title_full | Controllable Image Captioning with Feature Refinement and Multilayer Fusion |
title_fullStr | Controllable Image Captioning with Feature Refinement and Multilayer Fusion |
title_full_unstemmed | Controllable Image Captioning with Feature Refinement and Multilayer Fusion |
title_short | Controllable Image Captioning with Feature Refinement and Multilayer Fusion |
title_sort | controllable image captioning with feature refinement and multilayer fusion |
topic | controllable image captioning information-augmented embedding MR-WGCN similarity loss |
url | https://www.mdpi.com/2076-3417/13/8/5020 |
work_keys_str_mv | AT sendu controllableimagecaptioningwithfeaturerefinementandmultilayerfusion AT hongzhu controllableimagecaptioningwithfeaturerefinementandmultilayerfusion AT yujiazhang controllableimagecaptioningwithfeaturerefinementandmultilayerfusion AT dongwang controllableimagecaptioningwithfeaturerefinementandmultilayerfusion AT jingshi controllableimagecaptioningwithfeaturerefinementandmultilayerfusion AT nanxing controllableimagecaptioningwithfeaturerefinementandmultilayerfusion AT guangfenglin controllableimagecaptioningwithfeaturerefinementandmultilayerfusion AT huiyuzhou controllableimagecaptioningwithfeaturerefinementandmultilayerfusion |