Deep Liquid State Machines With Neural Plasticity for Video Activity Recognition

Real-world applications such as first-person video activity recognition require intelligent edge devices. However, size, weight, and power constraints of the embedded platforms cannot support resource intensive state-of-the-art algorithms. Machine learning lite algorithms, such as reservoir computin...

Full description

Bibliographic Details
Main Authors: Nicholas Soures, Dhireesha Kudithipudi
Format: Article
Language:English
Published: Frontiers Media S.A. 2019-07-01
Series:Frontiers in Neuroscience
Subjects:
Online Access:https://www.frontiersin.org/article/10.3389/fnins.2019.00686/full
_version_ 1819039658655350784
author Nicholas Soures
Dhireesha Kudithipudi
author_facet Nicholas Soures
Dhireesha Kudithipudi
author_sort Nicholas Soures
collection DOAJ
description Real-world applications such as first-person video activity recognition require intelligent edge devices. However, size, weight, and power constraints of the embedded platforms cannot support resource intensive state-of-the-art algorithms. Machine learning lite algorithms, such as reservoir computing, with shallow 3-layer networks are computationally frugal as only the output layer is trained. By reducing network depth and plasticity, reservoir computing minimizes computational power and complexity, making the algorithms optimal for edge devices. However, as a trade-off for their frugal nature, reservoir computing sacrifices computational power compared to state-of-the-art methods. A good compromise between reservoir computing and fully supervised networks are the proposed deep-LSM networks. The deep-LSM is a deep spiking neural network which captures dynamic information over multiple time-scales with a combination of randomly connected layers and unsupervised layers. The deep-LSM processes the captured dynamic information through an attention modulated readout layer to perform classification. We demonstrate that the deep-LSM achieves an average of 84.78% accuracy on the DogCentric video activity recognition task, beating state-of-the-art. The deep-LSM also shows up to 91.13% memory savings and up to 91.55% reduction in synaptic operations when compared to similar recurrent neural network models. Based on these results we claim that the deep-LSM is capable of overcoming limitations of traditional reservoir computing, while maintaining the low computational cost associated with reservoir computing.
first_indexed 2024-12-21T08:56:42Z
format Article
id doaj.art-624a28e815854be2b56c6df880b8b86f
institution Directory Open Access Journal
issn 1662-453X
language English
last_indexed 2024-12-21T08:56:42Z
publishDate 2019-07-01
publisher Frontiers Media S.A.
record_format Article
series Frontiers in Neuroscience
spelling doaj.art-624a28e815854be2b56c6df880b8b86f2022-12-21T19:09:32ZengFrontiers Media S.A.Frontiers in Neuroscience1662-453X2019-07-011310.3389/fnins.2019.00686457929Deep Liquid State Machines With Neural Plasticity for Video Activity RecognitionNicholas SouresDhireesha KudithipudiReal-world applications such as first-person video activity recognition require intelligent edge devices. However, size, weight, and power constraints of the embedded platforms cannot support resource intensive state-of-the-art algorithms. Machine learning lite algorithms, such as reservoir computing, with shallow 3-layer networks are computationally frugal as only the output layer is trained. By reducing network depth and plasticity, reservoir computing minimizes computational power and complexity, making the algorithms optimal for edge devices. However, as a trade-off for their frugal nature, reservoir computing sacrifices computational power compared to state-of-the-art methods. A good compromise between reservoir computing and fully supervised networks are the proposed deep-LSM networks. The deep-LSM is a deep spiking neural network which captures dynamic information over multiple time-scales with a combination of randomly connected layers and unsupervised layers. The deep-LSM processes the captured dynamic information through an attention modulated readout layer to perform classification. We demonstrate that the deep-LSM achieves an average of 84.78% accuracy on the DogCentric video activity recognition task, beating state-of-the-art. The deep-LSM also shows up to 91.13% memory savings and up to 91.55% reduction in synaptic operations when compared to similar recurrent neural network models. Based on these results we claim that the deep-LSM is capable of overcoming limitations of traditional reservoir computing, while maintaining the low computational cost associated with reservoir computing.https://www.frontiersin.org/article/10.3389/fnins.2019.00686/fullspikingLSMlocal learningdeeprecurrent
spellingShingle Nicholas Soures
Dhireesha Kudithipudi
Deep Liquid State Machines With Neural Plasticity for Video Activity Recognition
Frontiers in Neuroscience
spiking
LSM
local learning
deep
recurrent
title Deep Liquid State Machines With Neural Plasticity for Video Activity Recognition
title_full Deep Liquid State Machines With Neural Plasticity for Video Activity Recognition
title_fullStr Deep Liquid State Machines With Neural Plasticity for Video Activity Recognition
title_full_unstemmed Deep Liquid State Machines With Neural Plasticity for Video Activity Recognition
title_short Deep Liquid State Machines With Neural Plasticity for Video Activity Recognition
title_sort deep liquid state machines with neural plasticity for video activity recognition
topic spiking
LSM
local learning
deep
recurrent
url https://www.frontiersin.org/article/10.3389/fnins.2019.00686/full
work_keys_str_mv AT nicholassoures deepliquidstatemachineswithneuralplasticityforvideoactivityrecognition
AT dhireeshakudithipudi deepliquidstatemachineswithneuralplasticityforvideoactivityrecognition