Toward Improving Ensemble-Based Collaborative Inference at the Edge

Ensemble-based collaborative inference systems, Edge Ensembles, are deep learning edge inference systems that enhance accuracy by aggregating predictions from models deployed on each device. They offer several advantages, including scalability based on task complexity and decentralized functionality...

Full description

Bibliographic Details
Main Authors: Shungo Kumazawa, Jaehoon Yu, Kazushi Kawamura, Thiem Van Chu, Masato Motomura
Format: Article
Language:English
Published: IEEE 2024-01-01
Series:IEEE Access
Subjects:
Online Access:https://ieeexplore.ieee.org/document/10384371/
_version_ 1797243489925201920
author Shungo Kumazawa
Jaehoon Yu
Kazushi Kawamura
Thiem Van Chu
Masato Motomura
author_facet Shungo Kumazawa
Jaehoon Yu
Kazushi Kawamura
Thiem Van Chu
Masato Motomura
author_sort Shungo Kumazawa
collection DOAJ
description Ensemble-based collaborative inference systems, Edge Ensembles, are deep learning edge inference systems that enhance accuracy by aggregating predictions from models deployed on each device. They offer several advantages, including scalability based on task complexity and decentralized functionality without dependency on centralized servers. In general, ensemble methods effectively improve the accuracy of deep learning, and conventional research uses several model integration techniques for deep learning ensembles. Some of these existing integration methods are more effective than those used in previous Edge Ensembles. However, it remains uncertain whether these methods can be directly applied in the context of cooperative inference systems involving multiple edge devices. This study investigates the effectiveness of conventional model integration techniques, including cascade, weighted averaging, and test-time augmentation (TTA), when applied to Edge Ensembles to enhance their performance. Furthermore, we propose enhancements of these techniques tailored for Edge Ensembles. The cascade reduces the number of models required for inference but worsens latency by sequential inference processing. To address this latency issue, we propose <inline-formula> <tex-math notation="LaTeX">$m$ </tex-math></inline-formula>-parallel cascade, which adjusts the number of models processed simultaneously to <inline-formula> <tex-math notation="LaTeX">$m$ </tex-math></inline-formula>. We also propose learning TTA policies and weights for weighted averaging using ensemble prediction labels instead of ground truth labels. In the experiments, we verified the effectiveness of each technique for Edge Ensembles. The proposed <inline-formula> <tex-math notation="LaTeX">$m$ </tex-math></inline-formula>-parallel cascade achieved a 2.8 times reduction in latency compared to the conventional cascade, even with a 1.06 times increase in computational costs. Additionally, the ensemble label-based learning demonstrated comparable effectiveness to the approach using ground truth labels.
first_indexed 2024-04-24T18:55:56Z
format Article
id doaj.art-25f3b7022f904ef1b16610bfb7796aa4
institution Directory Open Access Journal
issn 2169-3536
language English
last_indexed 2024-04-24T18:55:56Z
publishDate 2024-01-01
publisher IEEE
record_format Article
series IEEE Access
spelling doaj.art-25f3b7022f904ef1b16610bfb7796aa42024-03-26T17:34:53ZengIEEEIEEE Access2169-35362024-01-01126926694010.1109/ACCESS.2024.335130810384371Toward Improving Ensemble-Based Collaborative Inference at the EdgeShungo Kumazawa0https://orcid.org/0009-0006-5608-5666Jaehoon Yu1https://orcid.org/0000-0001-6639-7694Kazushi Kawamura2https://orcid.org/0000-0002-0795-2974Thiem Van Chu3Masato Motomura4https://orcid.org/0000-0003-1543-1252Tokyo Institute of Technology, Midori-ku, Yokohama, JapanTokyo Institute of Technology, Midori-ku, Yokohama, JapanTokyo Institute of Technology, Midori-ku, Yokohama, JapanTokyo Institute of Technology, Midori-ku, Yokohama, JapanTokyo Institute of Technology, Midori-ku, Yokohama, JapanEnsemble-based collaborative inference systems, Edge Ensembles, are deep learning edge inference systems that enhance accuracy by aggregating predictions from models deployed on each device. They offer several advantages, including scalability based on task complexity and decentralized functionality without dependency on centralized servers. In general, ensemble methods effectively improve the accuracy of deep learning, and conventional research uses several model integration techniques for deep learning ensembles. Some of these existing integration methods are more effective than those used in previous Edge Ensembles. However, it remains uncertain whether these methods can be directly applied in the context of cooperative inference systems involving multiple edge devices. This study investigates the effectiveness of conventional model integration techniques, including cascade, weighted averaging, and test-time augmentation (TTA), when applied to Edge Ensembles to enhance their performance. Furthermore, we propose enhancements of these techniques tailored for Edge Ensembles. The cascade reduces the number of models required for inference but worsens latency by sequential inference processing. To address this latency issue, we propose <inline-formula> <tex-math notation="LaTeX">$m$ </tex-math></inline-formula>-parallel cascade, which adjusts the number of models processed simultaneously to <inline-formula> <tex-math notation="LaTeX">$m$ </tex-math></inline-formula>. We also propose learning TTA policies and weights for weighted averaging using ensemble prediction labels instead of ground truth labels. In the experiments, we verified the effectiveness of each technique for Edge Ensembles. The proposed <inline-formula> <tex-math notation="LaTeX">$m$ </tex-math></inline-formula>-parallel cascade achieved a 2.8 times reduction in latency compared to the conventional cascade, even with a 1.06 times increase in computational costs. Additionally, the ensemble label-based learning demonstrated comparable effectiveness to the approach using ground truth labels.https://ieeexplore.ieee.org/document/10384371/Ensembleedge computingcollaborative inferenceneural networkscascadetest time augmentation
spellingShingle Shungo Kumazawa
Jaehoon Yu
Kazushi Kawamura
Thiem Van Chu
Masato Motomura
Toward Improving Ensemble-Based Collaborative Inference at the Edge
IEEE Access
Ensemble
edge computing
collaborative inference
neural networks
cascade
test time augmentation
title Toward Improving Ensemble-Based Collaborative Inference at the Edge
title_full Toward Improving Ensemble-Based Collaborative Inference at the Edge
title_fullStr Toward Improving Ensemble-Based Collaborative Inference at the Edge
title_full_unstemmed Toward Improving Ensemble-Based Collaborative Inference at the Edge
title_short Toward Improving Ensemble-Based Collaborative Inference at the Edge
title_sort toward improving ensemble based collaborative inference at the edge
topic Ensemble
edge computing
collaborative inference
neural networks
cascade
test time augmentation
url https://ieeexplore.ieee.org/document/10384371/
work_keys_str_mv AT shungokumazawa towardimprovingensemblebasedcollaborativeinferenceattheedge
AT jaehoonyu towardimprovingensemblebasedcollaborativeinferenceattheedge
AT kazushikawamura towardimprovingensemblebasedcollaborativeinferenceattheedge
AT thiemvanchu towardimprovingensemblebasedcollaborativeinferenceattheedge
AT masatomotomura towardimprovingensemblebasedcollaborativeinferenceattheedge