A Study on the Application of TensorFlow Compression Techniques to Human Activity Recognition

In the human activity recognition (HAR) application domain, the use of deep learning (DL) algorithms for feature extractions and training purposes delivers significant performance improvements with respect to the use of traditional machine learning (ML) algorithms. However, this comes at the expense...

Full description

Bibliographic Details
Main Authors: Chiara Contoli, Emanuele Lattanzi
Format: Article
Language:English
Published: IEEE 2023-01-01
Series:IEEE Access
Subjects:
Online Access:https://ieeexplore.ieee.org/document/10124768/
_version_ 1797822229305622528
author Chiara Contoli
Emanuele Lattanzi
author_facet Chiara Contoli
Emanuele Lattanzi
author_sort Chiara Contoli
collection DOAJ
description In the human activity recognition (HAR) application domain, the use of deep learning (DL) algorithms for feature extractions and training purposes delivers significant performance improvements with respect to the use of traditional machine learning (ML) algorithms. However, this comes at the expense of more complex and demanding models, making harder their deployment on constrained devices traditionally involved in the HAR process. The efficiency of DL deployment is thus yet to be explored. We thoroughly investigated the application of TensorFlow Lite simple conversion, dynamic, and full integer quantization compression techniques. We applied those techniques not only to convolutional neural networks (CNNs), but also to long short-term memory (LSTM) networks, and a combined version of CNN and LSTM. We also considered two use case scenarios, namely cascading compression and stand-alone compression mode. This paper reports the feasibility of deploying deep networks onto an ESP32 device, and how TensorFlow compression techniques impact classification accuracy, energy consumption, and inference latency. Results show that in the cascading case, it is not possible to carry out the performance characterization. Whereas in the stand-alone case, dynamic quantization is recommended because yields a negligible loss of accuracy. In terms of power efficiency, both dynamic and full integer quantization provide high energy saving with respect to the uncompressed models: between 31% and 37% for CNN networks, and up to 45% for LSTM networks. In terms of inference latency, dynamic and full integer quantization provide comparable performance.
first_indexed 2024-03-13T10:04:58Z
format Article
id doaj.art-f87cb8f2ca3b4810b7a0d1bb56ae70e8
institution Directory Open Access Journal
issn 2169-3536
language English
last_indexed 2024-03-13T10:04:58Z
publishDate 2023-01-01
publisher IEEE
record_format Article
series IEEE Access
spelling doaj.art-f87cb8f2ca3b4810b7a0d1bb56ae70e82023-05-22T23:00:28ZengIEEEIEEE Access2169-35362023-01-0111480464805810.1109/ACCESS.2023.327643810124768A Study on the Application of TensorFlow Compression Techniques to Human Activity RecognitionChiara Contoli0https://orcid.org/0000-0003-2389-2593Emanuele Lattanzi1https://orcid.org/0000-0002-6568-8470Department of Pure and Applied Sciences, University of Urbino, Urbino, ItalyDepartment of Pure and Applied Sciences, University of Urbino, Urbino, ItalyIn the human activity recognition (HAR) application domain, the use of deep learning (DL) algorithms for feature extractions and training purposes delivers significant performance improvements with respect to the use of traditional machine learning (ML) algorithms. However, this comes at the expense of more complex and demanding models, making harder their deployment on constrained devices traditionally involved in the HAR process. The efficiency of DL deployment is thus yet to be explored. We thoroughly investigated the application of TensorFlow Lite simple conversion, dynamic, and full integer quantization compression techniques. We applied those techniques not only to convolutional neural networks (CNNs), but also to long short-term memory (LSTM) networks, and a combined version of CNN and LSTM. We also considered two use case scenarios, namely cascading compression and stand-alone compression mode. This paper reports the feasibility of deploying deep networks onto an ESP32 device, and how TensorFlow compression techniques impact classification accuracy, energy consumption, and inference latency. Results show that in the cascading case, it is not possible to carry out the performance characterization. Whereas in the stand-alone case, dynamic quantization is recommended because yields a negligible loss of accuracy. In terms of power efficiency, both dynamic and full integer quantization provide high energy saving with respect to the uncompressed models: between 31% and 37% for CNN networks, and up to 45% for LSTM networks. In terms of inference latency, dynamic and full integer quantization provide comparable performance.https://ieeexplore.ieee.org/document/10124768/Compression techniquesdeep learningdynamic quantizationESP32full integer quantizationhuman activity recognition
spellingShingle Chiara Contoli
Emanuele Lattanzi
A Study on the Application of TensorFlow Compression Techniques to Human Activity Recognition
IEEE Access
Compression techniques
deep learning
dynamic quantization
ESP32
full integer quantization
human activity recognition
title A Study on the Application of TensorFlow Compression Techniques to Human Activity Recognition
title_full A Study on the Application of TensorFlow Compression Techniques to Human Activity Recognition
title_fullStr A Study on the Application of TensorFlow Compression Techniques to Human Activity Recognition
title_full_unstemmed A Study on the Application of TensorFlow Compression Techniques to Human Activity Recognition
title_short A Study on the Application of TensorFlow Compression Techniques to Human Activity Recognition
title_sort study on the application of tensorflow compression techniques to human activity recognition
topic Compression techniques
deep learning
dynamic quantization
ESP32
full integer quantization
human activity recognition
url https://ieeexplore.ieee.org/document/10124768/
work_keys_str_mv AT chiaracontoli astudyontheapplicationoftensorflowcompressiontechniquestohumanactivityrecognition
AT emanuelelattanzi astudyontheapplicationoftensorflowcompressiontechniquestohumanactivityrecognition
AT chiaracontoli studyontheapplicationoftensorflowcompressiontechniquestohumanactivityrecognition
AT emanuelelattanzi studyontheapplicationoftensorflowcompressiontechniquestohumanactivityrecognition