Efficient inference offloading for mixture-of-experts large language models in internet of medical things
Despite recent significant advancements in large language models (LLMs) for medical services, the deployment difficulties of LLMs in e-healthcare hinder complex medical applications in the Internet of Medical Things (IoMT). People are increasingly concerned about e-healthcare risks and privacy prote...
Main Authors: | Yuan, Xiaoming, Kong, Weixuan, Luo, Zhenyu, Xu, Minrui |
---|---|
Other Authors: | School of Computer Science and Engineering |
Format: | Journal Article |
Language: | English |
Published: |
2024
|
Subjects: | |
Online Access: | https://hdl.handle.net/10356/179743 |
Similar Items
-
Computation offloading and content caching and delivery in Vehicular Edge Network: a survey
by: Dziyauddin, Rudzidatul Akmam, et al.
Published: (2022) -
Visible light based occupancy inference using ensemble learning
by: Hao, Jie, et al.
Published: (2018) -
Towards formal verification of Bayesian inference in probabilistic programming via guaranteed bounds
by: Zaiser, F
Published: (2024) -
Internet of Things: Principles and Paradigms /
by: Buyya, Rajkumar, editor 184757, et al.
Published: (2016) -
Arbitrarily strong utility-privacy tradeoff in multi-agent systems
by: Wang, Chong Xiao, et al.
Published: (2021)