Performance and Information Leakage in Splitfed Learning and Multi-Head Split Learning in Healthcare Data and Beyond
Machine learning (ML) in healthcare data analytics is attracting much attention because of the unprecedented power of ML to extract knowledge that improves the decision-making process. At the same time, laws and ethics codes drafted by countries to govern healthcare data are becoming stringent. Alth...
Main Authors: | , , , , , |
---|---|
Format: | Article |
Language: | English |
Published: |
MDPI AG
2022-07-01
|
Series: | Methods and Protocols |
Subjects: | |
Online Access: | https://www.mdpi.com/2409-9279/5/4/60 |
_version_ | 1797408627340869632 |
---|---|
author | Praveen Joshi Chandra Thapa Seyit Camtepe Mohammed Hasanuzzaman Ted Scully Haithem Afli |
author_facet | Praveen Joshi Chandra Thapa Seyit Camtepe Mohammed Hasanuzzaman Ted Scully Haithem Afli |
author_sort | Praveen Joshi |
collection | DOAJ |
description | Machine learning (ML) in healthcare data analytics is attracting much attention because of the unprecedented power of ML to extract knowledge that improves the decision-making process. At the same time, laws and ethics codes drafted by countries to govern healthcare data are becoming stringent. Although healthcare practitioners are struggling with an enforced governance framework, we see the emergence of distributed learning-based frameworks disrupting traditional-ML-model development. Splitfed learning (SFL) is one of the recent developments in distributed machine learning that empowers healthcare practitioners to preserve the privacy of input data and enables them to train ML models. However, SFL has some extra communication and computation overheads at the client side due to the requirement of client-side model synchronization. For a resource-constrained client side (hospitals with limited computational powers), removing such conditions is required to gain efficiency in the learning. In this regard, this paper studies SFL without client-side model synchronization. The resulting architecture is known as multi-head split learning (MHSL). At the same time, it is important to investigate information leakage, which indicates how much information is gained by the server related to the raw data directly out of the smashed data—the output of the client-side model portion—passed to it by the client. Our empirical studies examine the Resnet-18 and Conv1-D architecture model on the ECG and HAM-10000 datasets under IID data distribution. The results find that SFL provides 1.81% and 2.36% better accuracy than MHSL on the ECG and HAM-10000 datasets, respectively (for cut-layer value set to 1). Analysis of experimentation with various client-side model portions demonstrates that it has an impact on the overall performance. With an increase in layers in the client-side model portion, SFL performance improves while MHSL performance degrades. Experiment results also demonstrate that information leakage provided by mutual information score values in SFL is more than MHSL for ECG and HAM-10000 datasets by <inline-formula><math xmlns="http://www.w3.org/1998/Math/MathML" display="inline"><semantics><mrow><mn>2</mn><mo>×</mo><msup><mn>10</mn><mrow><mo>−</mo><mn>5</mn></mrow></msup></mrow></semantics></math></inline-formula> and <inline-formula><math xmlns="http://www.w3.org/1998/Math/MathML" display="inline"><semantics><mrow><mn>4</mn><mo>×</mo><msup><mn>10</mn><mrow><mo>−</mo><mn>3</mn></mrow></msup></mrow></semantics></math></inline-formula>, respectively. |
first_indexed | 2024-03-09T04:01:30Z |
format | Article |
id | doaj.art-6baad551885c4865b5cb3a6f337d5ff7 |
institution | Directory Open Access Journal |
issn | 2409-9279 |
language | English |
last_indexed | 2024-03-09T04:01:30Z |
publishDate | 2022-07-01 |
publisher | MDPI AG |
record_format | Article |
series | Methods and Protocols |
spelling | doaj.art-6baad551885c4865b5cb3a6f337d5ff72023-12-03T14:12:31ZengMDPI AGMethods and Protocols2409-92792022-07-01546010.3390/mps5040060Performance and Information Leakage in Splitfed Learning and Multi-Head Split Learning in Healthcare Data and BeyondPraveen Joshi0Chandra Thapa1Seyit Camtepe2Mohammed Hasanuzzaman3Ted Scully4Haithem Afli5Department of Computer Sciences, Munster Technological University, MTU, T12 P928 Cork, IrelandCSIRO Data61, Marsfield, NSW 2122, AustraliaCSIRO Data61, Marsfield, NSW 2122, AustraliaDepartment of Computer Sciences, Munster Technological University, MTU, T12 P928 Cork, IrelandDepartment of Computer Sciences, Munster Technological University, MTU, T12 P928 Cork, IrelandDepartment of Computer Sciences, Munster Technological University, MTU, T12 P928 Cork, IrelandMachine learning (ML) in healthcare data analytics is attracting much attention because of the unprecedented power of ML to extract knowledge that improves the decision-making process. At the same time, laws and ethics codes drafted by countries to govern healthcare data are becoming stringent. Although healthcare practitioners are struggling with an enforced governance framework, we see the emergence of distributed learning-based frameworks disrupting traditional-ML-model development. Splitfed learning (SFL) is one of the recent developments in distributed machine learning that empowers healthcare practitioners to preserve the privacy of input data and enables them to train ML models. However, SFL has some extra communication and computation overheads at the client side due to the requirement of client-side model synchronization. For a resource-constrained client side (hospitals with limited computational powers), removing such conditions is required to gain efficiency in the learning. In this regard, this paper studies SFL without client-side model synchronization. The resulting architecture is known as multi-head split learning (MHSL). At the same time, it is important to investigate information leakage, which indicates how much information is gained by the server related to the raw data directly out of the smashed data—the output of the client-side model portion—passed to it by the client. Our empirical studies examine the Resnet-18 and Conv1-D architecture model on the ECG and HAM-10000 datasets under IID data distribution. The results find that SFL provides 1.81% and 2.36% better accuracy than MHSL on the ECG and HAM-10000 datasets, respectively (for cut-layer value set to 1). Analysis of experimentation with various client-side model portions demonstrates that it has an impact on the overall performance. With an increase in layers in the client-side model portion, SFL performance improves while MHSL performance degrades. Experiment results also demonstrate that information leakage provided by mutual information score values in SFL is more than MHSL for ECG and HAM-10000 datasets by <inline-formula><math xmlns="http://www.w3.org/1998/Math/MathML" display="inline"><semantics><mrow><mn>2</mn><mo>×</mo><msup><mn>10</mn><mrow><mo>−</mo><mn>5</mn></mrow></msup></mrow></semantics></math></inline-formula> and <inline-formula><math xmlns="http://www.w3.org/1998/Math/MathML" display="inline"><semantics><mrow><mn>4</mn><mo>×</mo><msup><mn>10</mn><mrow><mo>−</mo><mn>3</mn></mrow></msup></mrow></semantics></math></inline-formula>, respectively.https://www.mdpi.com/2409-9279/5/4/60distributed collaborative machine learningsplit learningmulti-head split learningparameter transmission-based distributed machine learningprivacy-preserving machine learninginformation leakage in distributed learning |
spellingShingle | Praveen Joshi Chandra Thapa Seyit Camtepe Mohammed Hasanuzzaman Ted Scully Haithem Afli Performance and Information Leakage in Splitfed Learning and Multi-Head Split Learning in Healthcare Data and Beyond Methods and Protocols distributed collaborative machine learning split learning multi-head split learning parameter transmission-based distributed machine learning privacy-preserving machine learning information leakage in distributed learning |
title | Performance and Information Leakage in Splitfed Learning and Multi-Head Split Learning in Healthcare Data and Beyond |
title_full | Performance and Information Leakage in Splitfed Learning and Multi-Head Split Learning in Healthcare Data and Beyond |
title_fullStr | Performance and Information Leakage in Splitfed Learning and Multi-Head Split Learning in Healthcare Data and Beyond |
title_full_unstemmed | Performance and Information Leakage in Splitfed Learning and Multi-Head Split Learning in Healthcare Data and Beyond |
title_short | Performance and Information Leakage in Splitfed Learning and Multi-Head Split Learning in Healthcare Data and Beyond |
title_sort | performance and information leakage in splitfed learning and multi head split learning in healthcare data and beyond |
topic | distributed collaborative machine learning split learning multi-head split learning parameter transmission-based distributed machine learning privacy-preserving machine learning information leakage in distributed learning |
url | https://www.mdpi.com/2409-9279/5/4/60 |
work_keys_str_mv | AT praveenjoshi performanceandinformationleakageinsplitfedlearningandmultiheadsplitlearninginhealthcaredataandbeyond AT chandrathapa performanceandinformationleakageinsplitfedlearningandmultiheadsplitlearninginhealthcaredataandbeyond AT seyitcamtepe performanceandinformationleakageinsplitfedlearningandmultiheadsplitlearninginhealthcaredataandbeyond AT mohammedhasanuzzaman performanceandinformationleakageinsplitfedlearningandmultiheadsplitlearninginhealthcaredataandbeyond AT tedscully performanceandinformationleakageinsplitfedlearningandmultiheadsplitlearninginhealthcaredataandbeyond AT haithemafli performanceandinformationleakageinsplitfedlearningandmultiheadsplitlearninginhealthcaredataandbeyond |