Inducing high energy-latency of large vision-language models with verbose images
Large vision-language models (VLMs) such as GPT-4 have achieved exceptional performance across various multi-modal tasks. However, the deployment of VLMs necessitates substantial energy consumption and computational resources. Once attackers maliciously induce high energy consumption and latency tim...
Main Authors: | Gao, K, Bai, Y, Gu, J, Xia, ST, Torr, P, Li, Z, Liu, W |
---|---|
Format: | Conference item |
Language: | English |
Published: |
OpenReview
2024
|
Similar Items
-
Energy-latency manipulation of multi-modal large language models via verbose samples
by: Gao, K, et al.
Published: (2024) -
The verbosity epidemic.
by: Grais, R, et al.
Published: (2008) -
The Role of Inhibition in Age-Related Off-Topic Verbosity: Not Access but Deletion and Restraint Functions
by: Shufei eYin, et al.
Published: (2016-04-01) -
Review of Verbosity and Description in the Sahifeh Sajjadieh Translation according to Antoine Berman Theory (Case Study: Ansaryan’s Translation)
by: mohamad farhadei, et al.
Published: (2017-09-01) -
LAVT: Language-Aware Vision Transformer for referring image segmentation
by: Yang, Z, et al.
Published: (2022)