A Study on the Application of TensorFlow Compression Techniques to Human Activity Recognition
In the human activity recognition (HAR) application domain, the use of deep learning (DL) algorithms for feature extractions and training purposes delivers significant performance improvements with respect to the use of traditional machine learning (ML) algorithms. However, this comes at the expense...
Main Authors: | , |
---|---|
Format: | Article |
Language: | English |
Published: |
IEEE
2023-01-01
|
Series: | IEEE Access |
Subjects: | |
Online Access: | https://ieeexplore.ieee.org/document/10124768/ |
_version_ | 1797822229305622528 |
---|---|
author | Chiara Contoli Emanuele Lattanzi |
author_facet | Chiara Contoli Emanuele Lattanzi |
author_sort | Chiara Contoli |
collection | DOAJ |
description | In the human activity recognition (HAR) application domain, the use of deep learning (DL) algorithms for feature extractions and training purposes delivers significant performance improvements with respect to the use of traditional machine learning (ML) algorithms. However, this comes at the expense of more complex and demanding models, making harder their deployment on constrained devices traditionally involved in the HAR process. The efficiency of DL deployment is thus yet to be explored. We thoroughly investigated the application of TensorFlow Lite simple conversion, dynamic, and full integer quantization compression techniques. We applied those techniques not only to convolutional neural networks (CNNs), but also to long short-term memory (LSTM) networks, and a combined version of CNN and LSTM. We also considered two use case scenarios, namely cascading compression and stand-alone compression mode. This paper reports the feasibility of deploying deep networks onto an ESP32 device, and how TensorFlow compression techniques impact classification accuracy, energy consumption, and inference latency. Results show that in the cascading case, it is not possible to carry out the performance characterization. Whereas in the stand-alone case, dynamic quantization is recommended because yields a negligible loss of accuracy. In terms of power efficiency, both dynamic and full integer quantization provide high energy saving with respect to the uncompressed models: between 31% and 37% for CNN networks, and up to 45% for LSTM networks. In terms of inference latency, dynamic and full integer quantization provide comparable performance. |
first_indexed | 2024-03-13T10:04:58Z |
format | Article |
id | doaj.art-f87cb8f2ca3b4810b7a0d1bb56ae70e8 |
institution | Directory Open Access Journal |
issn | 2169-3536 |
language | English |
last_indexed | 2024-03-13T10:04:58Z |
publishDate | 2023-01-01 |
publisher | IEEE |
record_format | Article |
series | IEEE Access |
spelling | doaj.art-f87cb8f2ca3b4810b7a0d1bb56ae70e82023-05-22T23:00:28ZengIEEEIEEE Access2169-35362023-01-0111480464805810.1109/ACCESS.2023.327643810124768A Study on the Application of TensorFlow Compression Techniques to Human Activity RecognitionChiara Contoli0https://orcid.org/0000-0003-2389-2593Emanuele Lattanzi1https://orcid.org/0000-0002-6568-8470Department of Pure and Applied Sciences, University of Urbino, Urbino, ItalyDepartment of Pure and Applied Sciences, University of Urbino, Urbino, ItalyIn the human activity recognition (HAR) application domain, the use of deep learning (DL) algorithms for feature extractions and training purposes delivers significant performance improvements with respect to the use of traditional machine learning (ML) algorithms. However, this comes at the expense of more complex and demanding models, making harder their deployment on constrained devices traditionally involved in the HAR process. The efficiency of DL deployment is thus yet to be explored. We thoroughly investigated the application of TensorFlow Lite simple conversion, dynamic, and full integer quantization compression techniques. We applied those techniques not only to convolutional neural networks (CNNs), but also to long short-term memory (LSTM) networks, and a combined version of CNN and LSTM. We also considered two use case scenarios, namely cascading compression and stand-alone compression mode. This paper reports the feasibility of deploying deep networks onto an ESP32 device, and how TensorFlow compression techniques impact classification accuracy, energy consumption, and inference latency. Results show that in the cascading case, it is not possible to carry out the performance characterization. Whereas in the stand-alone case, dynamic quantization is recommended because yields a negligible loss of accuracy. In terms of power efficiency, both dynamic and full integer quantization provide high energy saving with respect to the uncompressed models: between 31% and 37% for CNN networks, and up to 45% for LSTM networks. In terms of inference latency, dynamic and full integer quantization provide comparable performance.https://ieeexplore.ieee.org/document/10124768/Compression techniquesdeep learningdynamic quantizationESP32full integer quantizationhuman activity recognition |
spellingShingle | Chiara Contoli Emanuele Lattanzi A Study on the Application of TensorFlow Compression Techniques to Human Activity Recognition IEEE Access Compression techniques deep learning dynamic quantization ESP32 full integer quantization human activity recognition |
title | A Study on the Application of TensorFlow Compression Techniques to Human Activity Recognition |
title_full | A Study on the Application of TensorFlow Compression Techniques to Human Activity Recognition |
title_fullStr | A Study on the Application of TensorFlow Compression Techniques to Human Activity Recognition |
title_full_unstemmed | A Study on the Application of TensorFlow Compression Techniques to Human Activity Recognition |
title_short | A Study on the Application of TensorFlow Compression Techniques to Human Activity Recognition |
title_sort | study on the application of tensorflow compression techniques to human activity recognition |
topic | Compression techniques deep learning dynamic quantization ESP32 full integer quantization human activity recognition |
url | https://ieeexplore.ieee.org/document/10124768/ |
work_keys_str_mv | AT chiaracontoli astudyontheapplicationoftensorflowcompressiontechniquestohumanactivityrecognition AT emanuelelattanzi astudyontheapplicationoftensorflowcompressiontechniquestohumanactivityrecognition AT chiaracontoli studyontheapplicationoftensorflowcompressiontechniquestohumanactivityrecognition AT emanuelelattanzi studyontheapplicationoftensorflowcompressiontechniquestohumanactivityrecognition |