A medical multimodal large language model for future pandemics

Deep neural networks have been integrated into the whole clinical decision procedure which can improve the efficiency of diagnosis and alleviate the heavy workload of physicians. Since most neural networks are supervised, their performance heavily depends on the volume and quality of available label...

Täydet tiedot

Bibliografiset tiedot
Päätekijät: Liu, F, Zhu, T, Wu, X, Yang, B, You, C, Wang, C, Lu, L, Liu, Z, Zheng, Y, Sun, X, Yang, Y, Clifton, L, Clifton, DA
Aineistotyyppi: Journal article
Kieli:English
Julkaistu: Springer Nature 2023
_version_ 1826311779341303808
author Liu, F
Zhu, T
Wu, X
Yang, B
You, C
Wang, C
Lu, L
Liu, Z
Zheng, Y
Sun, X
Yang, Y
Clifton, L
Clifton, DA
author_facet Liu, F
Zhu, T
Wu, X
Yang, B
You, C
Wang, C
Lu, L
Liu, Z
Zheng, Y
Sun, X
Yang, Y
Clifton, L
Clifton, DA
author_sort Liu, F
collection OXFORD
description Deep neural networks have been integrated into the whole clinical decision procedure which can improve the efficiency of diagnosis and alleviate the heavy workload of physicians. Since most neural networks are supervised, their performance heavily depends on the volume and quality of available labels. However, few such labels exist for rare diseases (e.g., new pandemics). Here we report a medical multimodal large language model (Med-MLLM) for radiograph representation learning, which can learn broad medical knowledge (e.g., image understanding, text semantics, and clinical phenotypes) from unlabelled data. As a result, when encountering a rare disease, our Med-MLLM can be rapidly deployed and easily adapted to them with limited labels. Furthermore, our model supports medical data across visual modality (e.g., chest X-ray and CT) and textual modality (e.g., medical report and free-text clinical note); therefore, it can be used for clinical tasks that involve both visual and textual data. We demonstrate the effectiveness of our Med-MLLM by showing how it would perform using the COVID-19 pandemic “in replay”. In the retrospective setting, we test the model on the early COVID-19 datasets; and in the prospective setting, we test the model on the new variant COVID-19-Omicron. The experiments are conducted on 1) three kinds of input data; 2) three kinds of downstream tasks, including disease reporting, diagnosis, and prognosis; 3) five COVID-19 datasets; and 4) three different languages, including English, Chinese, and Spanish. All experiments show that our model can make accurate and robust COVID-19 decision-support with little labelled data.
first_indexed 2024-03-07T08:14:47Z
format Journal article
id oxford-uuid:dc358e68-1be1-4c7b-a478-7e0023fda5c5
institution University of Oxford
language English
last_indexed 2024-03-07T08:14:47Z
publishDate 2023
publisher Springer Nature
record_format dspace
spelling oxford-uuid:dc358e68-1be1-4c7b-a478-7e0023fda5c52023-12-19T11:31:52ZA medical multimodal large language model for future pandemicsJournal articlehttp://purl.org/coar/resource_type/c_dcae04bcuuid:dc358e68-1be1-4c7b-a478-7e0023fda5c5EnglishSymplectic ElementsSpringer Nature2023Liu, FZhu, TWu, XYang, BYou, CWang, CLu, LLiu, ZZheng, YSun, XYang, YClifton, LClifton, DADeep neural networks have been integrated into the whole clinical decision procedure which can improve the efficiency of diagnosis and alleviate the heavy workload of physicians. Since most neural networks are supervised, their performance heavily depends on the volume and quality of available labels. However, few such labels exist for rare diseases (e.g., new pandemics). Here we report a medical multimodal large language model (Med-MLLM) for radiograph representation learning, which can learn broad medical knowledge (e.g., image understanding, text semantics, and clinical phenotypes) from unlabelled data. As a result, when encountering a rare disease, our Med-MLLM can be rapidly deployed and easily adapted to them with limited labels. Furthermore, our model supports medical data across visual modality (e.g., chest X-ray and CT) and textual modality (e.g., medical report and free-text clinical note); therefore, it can be used for clinical tasks that involve both visual and textual data. We demonstrate the effectiveness of our Med-MLLM by showing how it would perform using the COVID-19 pandemic “in replay”. In the retrospective setting, we test the model on the early COVID-19 datasets; and in the prospective setting, we test the model on the new variant COVID-19-Omicron. The experiments are conducted on 1) three kinds of input data; 2) three kinds of downstream tasks, including disease reporting, diagnosis, and prognosis; 3) five COVID-19 datasets; and 4) three different languages, including English, Chinese, and Spanish. All experiments show that our model can make accurate and robust COVID-19 decision-support with little labelled data.
spellingShingle Liu, F
Zhu, T
Wu, X
Yang, B
You, C
Wang, C
Lu, L
Liu, Z
Zheng, Y
Sun, X
Yang, Y
Clifton, L
Clifton, DA
A medical multimodal large language model for future pandemics
title A medical multimodal large language model for future pandemics
title_full A medical multimodal large language model for future pandemics
title_fullStr A medical multimodal large language model for future pandemics
title_full_unstemmed A medical multimodal large language model for future pandemics
title_short A medical multimodal large language model for future pandemics
title_sort medical multimodal large language model for future pandemics
work_keys_str_mv AT liuf amedicalmultimodallargelanguagemodelforfuturepandemics
AT zhut amedicalmultimodallargelanguagemodelforfuturepandemics
AT wux amedicalmultimodallargelanguagemodelforfuturepandemics
AT yangb amedicalmultimodallargelanguagemodelforfuturepandemics
AT youc amedicalmultimodallargelanguagemodelforfuturepandemics
AT wangc amedicalmultimodallargelanguagemodelforfuturepandemics
AT lul amedicalmultimodallargelanguagemodelforfuturepandemics
AT liuz amedicalmultimodallargelanguagemodelforfuturepandemics
AT zhengy amedicalmultimodallargelanguagemodelforfuturepandemics
AT sunx amedicalmultimodallargelanguagemodelforfuturepandemics
AT yangy amedicalmultimodallargelanguagemodelforfuturepandemics
AT cliftonl amedicalmultimodallargelanguagemodelforfuturepandemics
AT cliftonda amedicalmultimodallargelanguagemodelforfuturepandemics
AT liuf medicalmultimodallargelanguagemodelforfuturepandemics
AT zhut medicalmultimodallargelanguagemodelforfuturepandemics
AT wux medicalmultimodallargelanguagemodelforfuturepandemics
AT yangb medicalmultimodallargelanguagemodelforfuturepandemics
AT youc medicalmultimodallargelanguagemodelforfuturepandemics
AT wangc medicalmultimodallargelanguagemodelforfuturepandemics
AT lul medicalmultimodallargelanguagemodelforfuturepandemics
AT liuz medicalmultimodallargelanguagemodelforfuturepandemics
AT zhengy medicalmultimodallargelanguagemodelforfuturepandemics
AT sunx medicalmultimodallargelanguagemodelforfuturepandemics
AT yangy medicalmultimodallargelanguagemodelforfuturepandemics
AT cliftonl medicalmultimodallargelanguagemodelforfuturepandemics
AT cliftonda medicalmultimodallargelanguagemodelforfuturepandemics