Dual-Dependency Attention Transformer for Fine-Grained Visual Classification
Visual transformers (ViTs) are widely used in various visual tasks, such as fine-grained visual classification (FGVC). However, the self-attention mechanism, which is the core module of visual transformers, leads to quadratic computational and memory complexity. The sparse-attention and local-attent...
Main Authors: | Shiyan Cui, Bin Hui |
---|---|
Format: | Article |
Language: | English |
Published: |
MDPI AG
2024-04-01
|
Series: | Sensors |
Subjects: | |
Online Access: | https://www.mdpi.com/1424-8220/24/7/2337 |
Similar Items
-
MFVT: Multilevel Feature Fusion Vision Transformer and RAMix Data Augmentation for Fine-Grained Visual Categorization
by: Xinyao Lv, et al.
Published: (2022-10-01) -
Token-Selective Vision Transformer for fine-grained image recognition of marine organisms
by: Guangzhe Si, et al.
Published: (2023-04-01) -
Loop and distillation: Attention weights fusion transformer for fine‐grained representation
by: Sun Fayou, et al.
Published: (2023-06-01) -
Hybrid Granularities Transformer for Fine-Grained Image Recognition
by: Ying Yu, et al.
Published: (2023-04-01) -
Few-shot fine-grained fish species classification via sandwich attention CovaMNet
by: Jiping Zhai, et al.
Published: (2023-03-01)