Static Video Summarization Using Video Coding Features with Frame-Level Temporal Subsampling and Deep Learning

There is an abundance of digital video content due to the cloud’s phenomenal growth and security footage; it is therefore essential to summarize these videos in data centers. This paper offers innovative approaches to the problem of key frame extraction for the purpose of video summarization. Our ap...

Full description

Bibliographic Details
Main Authors: Obada Issa, Tamer Shanableh
Format: Article
Language:English
Published: MDPI AG 2023-05-01
Series:Applied Sciences
Subjects:
Online Access:https://www.mdpi.com/2076-3417/13/10/6065
_version_ 1797601265889312768
author Obada Issa
Tamer Shanableh
author_facet Obada Issa
Tamer Shanableh
author_sort Obada Issa
collection DOAJ
description There is an abundance of digital video content due to the cloud’s phenomenal growth and security footage; it is therefore essential to summarize these videos in data centers. This paper offers innovative approaches to the problem of key frame extraction for the purpose of video summarization. Our approach includes the extraction of feature variables from the bit streams of coded videos, followed by optional stepwise regression for dimensionality reduction. Once the features are extracted and their dimensionality is reduced, we apply innovative frame-level temporal subsampling techniques, followed by training and testing using deep learning architectures. The frame-level temporal subsampling techniques are based on cosine similarity and the PCA projections of feature vectors. We create three different learning architectures by utilizing LSTM networks, 1D-CNN networks, and random forests. The four most popular video summarization datasets, namely, TVSum, SumMe, OVP, and VSUMM, are used to evaluate the accuracy of the proposed solutions. This includes the precision, recall, F-score measures, and computational time. It is shown that the proposed solutions, when trained and tested on all subjective user summaries, achieved F-scores of 0.79, 0.74, 0.88, and 0.81, respectively, for the aforementioned datasets, showing clear improvements over prior studies.
first_indexed 2024-03-11T03:58:43Z
format Article
id doaj.art-7e8414ce5c4d40e787fddb8daae27371
institution Directory Open Access Journal
issn 2076-3417
language English
last_indexed 2024-03-11T03:58:43Z
publishDate 2023-05-01
publisher MDPI AG
record_format Article
series Applied Sciences
spelling doaj.art-7e8414ce5c4d40e787fddb8daae273712023-11-18T00:20:00ZengMDPI AGApplied Sciences2076-34172023-05-011310606510.3390/app13106065Static Video Summarization Using Video Coding Features with Frame-Level Temporal Subsampling and Deep LearningObada Issa0Tamer Shanableh1Department of Computer Science and Engineering, American University of Sharjah, Sharjah P.O. Box 26666, United Arab EmiratesDepartment of Computer Science and Engineering, American University of Sharjah, Sharjah P.O. Box 26666, United Arab EmiratesThere is an abundance of digital video content due to the cloud’s phenomenal growth and security footage; it is therefore essential to summarize these videos in data centers. This paper offers innovative approaches to the problem of key frame extraction for the purpose of video summarization. Our approach includes the extraction of feature variables from the bit streams of coded videos, followed by optional stepwise regression for dimensionality reduction. Once the features are extracted and their dimensionality is reduced, we apply innovative frame-level temporal subsampling techniques, followed by training and testing using deep learning architectures. The frame-level temporal subsampling techniques are based on cosine similarity and the PCA projections of feature vectors. We create three different learning architectures by utilizing LSTM networks, 1D-CNN networks, and random forests. The four most popular video summarization datasets, namely, TVSum, SumMe, OVP, and VSUMM, are used to evaluate the accuracy of the proposed solutions. This includes the precision, recall, F-score measures, and computational time. It is shown that the proposed solutions, when trained and tested on all subjective user summaries, achieved F-scores of 0.79, 0.74, 0.88, and 0.81, respectively, for the aforementioned datasets, showing clear improvements over prior studies.https://www.mdpi.com/2076-3417/13/10/6065video summarizationvideo codingtemporal subsamplingconvolution neural networkslong-short term memory
spellingShingle Obada Issa
Tamer Shanableh
Static Video Summarization Using Video Coding Features with Frame-Level Temporal Subsampling and Deep Learning
Applied Sciences
video summarization
video coding
temporal subsampling
convolution neural networks
long-short term memory
title Static Video Summarization Using Video Coding Features with Frame-Level Temporal Subsampling and Deep Learning
title_full Static Video Summarization Using Video Coding Features with Frame-Level Temporal Subsampling and Deep Learning
title_fullStr Static Video Summarization Using Video Coding Features with Frame-Level Temporal Subsampling and Deep Learning
title_full_unstemmed Static Video Summarization Using Video Coding Features with Frame-Level Temporal Subsampling and Deep Learning
title_short Static Video Summarization Using Video Coding Features with Frame-Level Temporal Subsampling and Deep Learning
title_sort static video summarization using video coding features with frame level temporal subsampling and deep learning
topic video summarization
video coding
temporal subsampling
convolution neural networks
long-short term memory
url https://www.mdpi.com/2076-3417/13/10/6065
work_keys_str_mv AT obadaissa staticvideosummarizationusingvideocodingfeatureswithframeleveltemporalsubsamplinganddeeplearning
AT tamershanableh staticvideosummarizationusingvideocodingfeatureswithframeleveltemporalsubsamplinganddeeplearning