Adversarial Attentive Multi-Modal Embedding Learning for Image-Text Matching
Matching the image and text with deep models has been extensively studied in recent years. Mining the correlation between image and text to learn effective multi-modal features is crucial for image-text matching. However, most existing approaches model the different types of correlation independentl...
Main Authors: | Kaimin Wei, Zhibo Zhou |
---|---|
Format: | Article |
Language: | English |
Published: |
IEEE
2020-01-01
|
Series: | IEEE Access |
Subjects: | |
Online Access: | https://ieeexplore.ieee.org/document/9097848/ |
Similar Items
-
Multi-Modal Memory Enhancement Attention Network for Image-Text Matching
by: Zhong Ji, et al.
Published: (2020-01-01) -
DA-GAN: Dual Attention Generative Adversarial Network for Cross-Modal Retrieval
by: Liewu Cai, et al.
Published: (2022-01-01) -
Bi-Modal Learning With Channel-Wise Attention for Multi-Label Image Classification
by: Peng Li, et al.
Published: (2020-01-01) -
Short Text Embedding Autoencoders With Attention-Based Neighborhood Preservation
by: Chao Wei, et al.
Published: (2020-01-01) -
A survey of generative adversarial networks and their application in text-to-image synthesis
by: Wu Zeng, et al.
Published: (2023-11-01)