Making Sense of Neuromorphic Event Data for Human Action Recognition

Neuromorphic vision sensors provide low power sensing and capture salient spatial-temporal events. The majority of the existing neuromorphic sensing work focus on object detection. However, since they only record the events, they provide an efficient signal domain for privacy aware surveillance task...

Full description

Bibliographic Details
Main Authors: Salah Al-Obaidi, Hiba Al-Khafaji, Charith Abhayaratne
Format: Article
Language:English
Published: IEEE 2021-01-01
Series:IEEE Access
Subjects:
Online Access:https://ieeexplore.ieee.org/document/9446050/
_version_ 1818899919088386048
author Salah Al-Obaidi
Hiba Al-Khafaji
Charith Abhayaratne
author_facet Salah Al-Obaidi
Hiba Al-Khafaji
Charith Abhayaratne
author_sort Salah Al-Obaidi
collection DOAJ
description Neuromorphic vision sensors provide low power sensing and capture salient spatial-temporal events. The majority of the existing neuromorphic sensing work focus on object detection. However, since they only record the events, they provide an efficient signal domain for privacy aware surveillance tasks. This paper explores how the neuromorphic vision sensor data streams can be analysed for human action recognition, which is a challenging application. The proposed method is based on handcrafted features. It consists of a pre-processing step for removing the noisy events followed by the extraction of handcrafted local and global feature vectors corresponding to the underlying human action. The local features are extracted considering a set of high-order descriptive statistics from the spatio-temporal events in a time window slice, while the global features are extracted by considering the frequencies of occurrences of the temporal event sequences. Then, low complexity classifiers, such as, support vector machines (SVM) and K-Nearest Neighbours (KNNs), are trained using these feature vectors. The proposed method evaluation uses three groups of datasets: Emulator-based, re-recording-based and native NVS-based. The proposed method has outperformed the existing methods in terms of human action recognition accuracy rates by 0.54%, 19.3%, and 25.61% for E-KTH, E-UCF11 and E-HMDB51 datasets, respectively. This paper also reports results for three further datasets: E-UCF50, R-UCF50, and N-Actions, which are reported for the first time for human action recognition on neuromorphic vision sensor domain.
first_indexed 2024-12-19T19:55:36Z
format Article
id doaj.art-26a3f2dc17dd409e9550ff7e5d065665
institution Directory Open Access Journal
issn 2169-3536
language English
last_indexed 2024-12-19T19:55:36Z
publishDate 2021-01-01
publisher IEEE
record_format Article
series IEEE Access
spelling doaj.art-26a3f2dc17dd409e9550ff7e5d0656652022-12-21T20:07:48ZengIEEEIEEE Access2169-35362021-01-019826868270010.1109/ACCESS.2021.30857089446050Making Sense of Neuromorphic Event Data for Human Action RecognitionSalah Al-Obaidi0https://orcid.org/0000-0003-4521-7193Hiba Al-Khafaji1https://orcid.org/0000-0002-4977-9284Charith Abhayaratne2https://orcid.org/0000-0002-2799-7395Department of Electronic and Electrical Engineering, The University of Sheffield, Sheffield, U.K.Department of Electronic and Electrical Engineering, The University of Sheffield, Sheffield, U.K.Department of Electronic and Electrical Engineering, The University of Sheffield, Sheffield, U.K.Neuromorphic vision sensors provide low power sensing and capture salient spatial-temporal events. The majority of the existing neuromorphic sensing work focus on object detection. However, since they only record the events, they provide an efficient signal domain for privacy aware surveillance tasks. This paper explores how the neuromorphic vision sensor data streams can be analysed for human action recognition, which is a challenging application. The proposed method is based on handcrafted features. It consists of a pre-processing step for removing the noisy events followed by the extraction of handcrafted local and global feature vectors corresponding to the underlying human action. The local features are extracted considering a set of high-order descriptive statistics from the spatio-temporal events in a time window slice, while the global features are extracted by considering the frequencies of occurrences of the temporal event sequences. Then, low complexity classifiers, such as, support vector machines (SVM) and K-Nearest Neighbours (KNNs), are trained using these feature vectors. The proposed method evaluation uses three groups of datasets: Emulator-based, re-recording-based and native NVS-based. The proposed method has outperformed the existing methods in terms of human action recognition accuracy rates by 0.54%, 19.3%, and 25.61% for E-KTH, E-UCF11 and E-HMDB51 datasets, respectively. This paper also reports results for three further datasets: E-UCF50, R-UCF50, and N-Actions, which are reported for the first time for human action recognition on neuromorphic vision sensor domain.https://ieeexplore.ieee.org/document/9446050/Neuromorphic vision sensing (NVS)event camerasdynamic vision sensing (DVS)human action recognition (HAR)local featuresglobal features
spellingShingle Salah Al-Obaidi
Hiba Al-Khafaji
Charith Abhayaratne
Making Sense of Neuromorphic Event Data for Human Action Recognition
IEEE Access
Neuromorphic vision sensing (NVS)
event cameras
dynamic vision sensing (DVS)
human action recognition (HAR)
local features
global features
title Making Sense of Neuromorphic Event Data for Human Action Recognition
title_full Making Sense of Neuromorphic Event Data for Human Action Recognition
title_fullStr Making Sense of Neuromorphic Event Data for Human Action Recognition
title_full_unstemmed Making Sense of Neuromorphic Event Data for Human Action Recognition
title_short Making Sense of Neuromorphic Event Data for Human Action Recognition
title_sort making sense of neuromorphic event data for human action recognition
topic Neuromorphic vision sensing (NVS)
event cameras
dynamic vision sensing (DVS)
human action recognition (HAR)
local features
global features
url https://ieeexplore.ieee.org/document/9446050/
work_keys_str_mv AT salahalobaidi makingsenseofneuromorphiceventdataforhumanactionrecognition
AT hibaalkhafaji makingsenseofneuromorphiceventdataforhumanactionrecognition
AT charithabhayaratne makingsenseofneuromorphiceventdataforhumanactionrecognition