Action recognition using attention-based spatio-temporal VLAD networks and adaptive video sequences optimization
Abstract In the field of human action recognition, it is a long-standing challenge to characterize the video-level spatio-temporal features effectively. This is attributable in part to the inability of CNN to model long-range temporal information, especially for actions that consist of multiple stag...
Główni autorzy: | , , |
---|---|
Format: | Artykuł |
Język: | English |
Wydane: |
Nature Portfolio
2024-10-01
|
Seria: | Scientific Reports |
Dostęp online: | https://doi.org/10.1038/s41598-024-75640-6 |
_version_ | 1826990883200827392 |
---|---|
author | Zhengkui Weng Xinmin Li Shoujian Xiong |
author_facet | Zhengkui Weng Xinmin Li Shoujian Xiong |
author_sort | Zhengkui Weng |
collection | DOAJ |
description | Abstract In the field of human action recognition, it is a long-standing challenge to characterize the video-level spatio-temporal features effectively. This is attributable in part to the inability of CNN to model long-range temporal information, especially for actions that consist of multiple staged behaviors. In this paper, a novel attention-based spatio-temporal VLAD network (AST-VLAD) with self-attention model is developed to aggregate the informative deep features across the video according to the adaptive deep feature selected. Moreover, an overall automatic approach to adaptive video sequences optimization (AVSO) is proposed through shot segmentation and dynamic weighted sampling, the AVSO increase in the proportion of action-related frames and eliminate the redundant intervals. Then, based on the optimized video, a self-attention model is introduced in AST-VLAD to modeling the intrinsic spatio-temporal relationship of deep features instead of solving the frame-level features in an average or max pooling manner. Extensive experiments are conducted on two public benchmarks-HMDB51 and UCF101 for evaluation. As compared to the existing frameworks, results show that the proposed approach performs better or as well in the accuracy of classification on both HMDB51 (73.1% ) and UCF101 (96.0%) datasets. |
first_indexed | 2025-02-18T08:27:10Z |
format | Article |
id | doaj.art-2f21925b5c6e468d8d85d0f55be81f3c |
institution | Directory Open Access Journal |
issn | 2045-2322 |
language | English |
last_indexed | 2025-02-18T08:27:10Z |
publishDate | 2024-10-01 |
publisher | Nature Portfolio |
record_format | Article |
series | Scientific Reports |
spelling | doaj.art-2f21925b5c6e468d8d85d0f55be81f3c2024-11-03T12:25:08ZengNature PortfolioScientific Reports2045-23222024-10-0114111710.1038/s41598-024-75640-6Action recognition using attention-based spatio-temporal VLAD networks and adaptive video sequences optimizationZhengkui Weng0Xinmin Li1Shoujian Xiong2School of Automation, Qingdao UniversitySchool of Mathematics & Statistics, Qingdao UniversityZhejiang Lancoo Technology Co., LtdAbstract In the field of human action recognition, it is a long-standing challenge to characterize the video-level spatio-temporal features effectively. This is attributable in part to the inability of CNN to model long-range temporal information, especially for actions that consist of multiple staged behaviors. In this paper, a novel attention-based spatio-temporal VLAD network (AST-VLAD) with self-attention model is developed to aggregate the informative deep features across the video according to the adaptive deep feature selected. Moreover, an overall automatic approach to adaptive video sequences optimization (AVSO) is proposed through shot segmentation and dynamic weighted sampling, the AVSO increase in the proportion of action-related frames and eliminate the redundant intervals. Then, based on the optimized video, a self-attention model is introduced in AST-VLAD to modeling the intrinsic spatio-temporal relationship of deep features instead of solving the frame-level features in an average or max pooling manner. Extensive experiments are conducted on two public benchmarks-HMDB51 and UCF101 for evaluation. As compared to the existing frameworks, results show that the proposed approach performs better or as well in the accuracy of classification on both HMDB51 (73.1% ) and UCF101 (96.0%) datasets.https://doi.org/10.1038/s41598-024-75640-6 |
spellingShingle | Zhengkui Weng Xinmin Li Shoujian Xiong Action recognition using attention-based spatio-temporal VLAD networks and adaptive video sequences optimization Scientific Reports |
title | Action recognition using attention-based spatio-temporal VLAD networks and adaptive video sequences optimization |
title_full | Action recognition using attention-based spatio-temporal VLAD networks and adaptive video sequences optimization |
title_fullStr | Action recognition using attention-based spatio-temporal VLAD networks and adaptive video sequences optimization |
title_full_unstemmed | Action recognition using attention-based spatio-temporal VLAD networks and adaptive video sequences optimization |
title_short | Action recognition using attention-based spatio-temporal VLAD networks and adaptive video sequences optimization |
title_sort | action recognition using attention based spatio temporal vlad networks and adaptive video sequences optimization |
url | https://doi.org/10.1038/s41598-024-75640-6 |
work_keys_str_mv | AT zhengkuiweng actionrecognitionusingattentionbasedspatiotemporalvladnetworksandadaptivevideosequencesoptimization AT xinminli actionrecognitionusingattentionbasedspatiotemporalvladnetworksandadaptivevideosequencesoptimization AT shoujianxiong actionrecognitionusingattentionbasedspatiotemporalvladnetworksandadaptivevideosequencesoptimization |