MERGE: A model for multi-input biomedical federated learning

Summary: Driven by the deep learning (DL) revolution, artificial intelligence (AI) has become a fundamental tool for many biomedical tasks, including analyzing and classifying diagnostic images. Imaging, however, is not the only source of information. Tabular data, such as personal and genomic data...

Full description

Bibliographic Details
Main Authors: Bruno Casella, Walter Riviera, Marco Aldinucci, Gloria Menegaz
Format: Article
Language:English
Published: Elsevier 2023-11-01
Series:Patterns
Subjects:
Online Access:http://www.sciencedirect.com/science/article/pii/S2666389923002404
_version_ 1797630611819593728
author Bruno Casella
Walter Riviera
Marco Aldinucci
Gloria Menegaz
author_facet Bruno Casella
Walter Riviera
Marco Aldinucci
Gloria Menegaz
author_sort Bruno Casella
collection DOAJ
description Summary: Driven by the deep learning (DL) revolution, artificial intelligence (AI) has become a fundamental tool for many biomedical tasks, including analyzing and classifying diagnostic images. Imaging, however, is not the only source of information. Tabular data, such as personal and genomic data and blood test results, are routinely collected but rarely considered in DL pipelines. Nevertheless, DL requires large datasets that often must be pooled from different institutions, raising non-trivial privacy concerns. Federated learning (FL) is a cooperative learning paradigm that aims to address these issues by moving models instead of data across different institutions. Here, we present a federated multi-input architecture using images and tabular data as a methodology to enhance model performance while preserving data privacy. We evaluated it on two showcases: the prognosis of COVID-19 and patients’ stratification in Alzheimer’s disease, providing evidence of enhanced accuracy and F1 scores against single-input models and improved generalizability against non-federated models. The bigger picture: Deep learning models must be trained with large datasets, which often requires pooling data from different sites and sources. In research fields dealing with sensitive information subject to data regulations, such as biomedical research, data pooling can generate concerns about data access and sharing across institutions, which can affect performance, energy consumption, privacy, and security. Federated learning is a cooperative learning paradigm that addresses such concerns by sharing models instead of data across different institutions.
first_indexed 2024-03-11T11:09:31Z
format Article
id doaj.art-785fe518d3564ca2aedf7a4b157242b2
institution Directory Open Access Journal
issn 2666-3899
language English
last_indexed 2024-03-11T11:09:31Z
publishDate 2023-11-01
publisher Elsevier
record_format Article
series Patterns
spelling doaj.art-785fe518d3564ca2aedf7a4b157242b22023-11-12T04:41:04ZengElsevierPatterns2666-38992023-11-01411100856MERGE: A model for multi-input biomedical federated learningBruno Casella0Walter Riviera1Marco Aldinucci2Gloria Menegaz3Department of Computer Science, University of Turin, 10149 Turin, ItalyDepartment of Computer Science, University of Verona, 37134 Verona, Italy; Corresponding authorDepartment of Computer Science, University of Turin, 10149 Turin, ItalyDepartment of Engineering for Innovation Medicine, University of Verona, 37134 Verona, ItalySummary: Driven by the deep learning (DL) revolution, artificial intelligence (AI) has become a fundamental tool for many biomedical tasks, including analyzing and classifying diagnostic images. Imaging, however, is not the only source of information. Tabular data, such as personal and genomic data and blood test results, are routinely collected but rarely considered in DL pipelines. Nevertheless, DL requires large datasets that often must be pooled from different institutions, raising non-trivial privacy concerns. Federated learning (FL) is a cooperative learning paradigm that aims to address these issues by moving models instead of data across different institutions. Here, we present a federated multi-input architecture using images and tabular data as a methodology to enhance model performance while preserving data privacy. We evaluated it on two showcases: the prognosis of COVID-19 and patients’ stratification in Alzheimer’s disease, providing evidence of enhanced accuracy and F1 scores against single-input models and improved generalizability against non-federated models. The bigger picture: Deep learning models must be trained with large datasets, which often requires pooling data from different sites and sources. In research fields dealing with sensitive information subject to data regulations, such as biomedical research, data pooling can generate concerns about data access and sharing across institutions, which can affect performance, energy consumption, privacy, and security. Federated learning is a cooperative learning paradigm that addresses such concerns by sharing models instead of data across different institutions.http://www.sciencedirect.com/science/article/pii/S2666389923002404DSML 2: Proof-of-Concept: Data science output has been formulated, implemented, and tested for one domain/problem
spellingShingle Bruno Casella
Walter Riviera
Marco Aldinucci
Gloria Menegaz
MERGE: A model for multi-input biomedical federated learning
Patterns
DSML 2: Proof-of-Concept: Data science output has been formulated, implemented, and tested for one domain/problem
title MERGE: A model for multi-input biomedical federated learning
title_full MERGE: A model for multi-input biomedical federated learning
title_fullStr MERGE: A model for multi-input biomedical federated learning
title_full_unstemmed MERGE: A model for multi-input biomedical federated learning
title_short MERGE: A model for multi-input biomedical federated learning
title_sort merge a model for multi input biomedical federated learning
topic DSML 2: Proof-of-Concept: Data science output has been formulated, implemented, and tested for one domain/problem
url http://www.sciencedirect.com/science/article/pii/S2666389923002404
work_keys_str_mv AT brunocasella mergeamodelformultiinputbiomedicalfederatedlearning
AT walterriviera mergeamodelformultiinputbiomedicalfederatedlearning
AT marcoaldinucci mergeamodelformultiinputbiomedicalfederatedlearning
AT gloriamenegaz mergeamodelformultiinputbiomedicalfederatedlearning