No-Reference Video Quality Assessment Using Multi-Pooled, Saliency Weighted Deep Features and Decision Fusion

With the constantly growing popularity of video-based services and applications, no-reference video quality assessment (NR-VQA) has become a very hot research topic. Over the years, many different approaches have been introduced in the literature to evaluate the perceptual quality of digital videos....

Full description

Bibliographic Details
Main Author: Domonkos Varga
Format: Article
Language:English
Published: MDPI AG 2022-03-01
Series:Sensors
Subjects:
Online Access:https://www.mdpi.com/1424-8220/22/6/2209
_version_ 1797442460748611584
author Domonkos Varga
author_facet Domonkos Varga
author_sort Domonkos Varga
collection DOAJ
description With the constantly growing popularity of video-based services and applications, no-reference video quality assessment (NR-VQA) has become a very hot research topic. Over the years, many different approaches have been introduced in the literature to evaluate the perceptual quality of digital videos. Due to the advent of large benchmark video quality assessment databases, deep learning has attracted a significant amount of attention in this field in recent years. This paper presents a novel, innovative deep learning-based approach for NR-VQA that relies on a set of in parallel pre-trained convolutional neural networks (CNN) to characterize versatitely the potential image and video distortions. Specifically, temporally pooled and saliency weighted video-level deep features are extracted with the help of a set of pre-trained CNNs and mapped onto perceptual quality scores independently from each other. Finally, the quality scores coming from the different regressors are fused together to obtain the perceptual quality of a given video sequence. Extensive experiments demonstrate that the proposed method sets a new state-of-the-art on two large benchmark video quality assessment databases with authentic distortions. Moreover, the presented results underline that the decision fusion of multiple deep architectures can significantly benefit NR-VQA.
first_indexed 2024-03-09T12:42:10Z
format Article
id doaj.art-b8313c6291104e38a45bb0865b92b410
institution Directory Open Access Journal
issn 1424-8220
language English
last_indexed 2024-03-09T12:42:10Z
publishDate 2022-03-01
publisher MDPI AG
record_format Article
series Sensors
spelling doaj.art-b8313c6291104e38a45bb0865b92b4102023-11-30T22:17:44ZengMDPI AGSensors1424-82202022-03-01226220910.3390/s22062209No-Reference Video Quality Assessment Using Multi-Pooled, Saliency Weighted Deep Features and Decision FusionDomonkos Varga0Ronin Institute, Montclair, NJ 07043, USAWith the constantly growing popularity of video-based services and applications, no-reference video quality assessment (NR-VQA) has become a very hot research topic. Over the years, many different approaches have been introduced in the literature to evaluate the perceptual quality of digital videos. Due to the advent of large benchmark video quality assessment databases, deep learning has attracted a significant amount of attention in this field in recent years. This paper presents a novel, innovative deep learning-based approach for NR-VQA that relies on a set of in parallel pre-trained convolutional neural networks (CNN) to characterize versatitely the potential image and video distortions. Specifically, temporally pooled and saliency weighted video-level deep features are extracted with the help of a set of pre-trained CNNs and mapped onto perceptual quality scores independently from each other. Finally, the quality scores coming from the different regressors are fused together to obtain the perceptual quality of a given video sequence. Extensive experiments demonstrate that the proposed method sets a new state-of-the-art on two large benchmark video quality assessment databases with authentic distortions. Moreover, the presented results underline that the decision fusion of multiple deep architectures can significantly benefit NR-VQA.https://www.mdpi.com/1424-8220/22/6/2209no-reference video quality assessmentconvolutional neural networkdecision fusion
spellingShingle Domonkos Varga
No-Reference Video Quality Assessment Using Multi-Pooled, Saliency Weighted Deep Features and Decision Fusion
Sensors
no-reference video quality assessment
convolutional neural network
decision fusion
title No-Reference Video Quality Assessment Using Multi-Pooled, Saliency Weighted Deep Features and Decision Fusion
title_full No-Reference Video Quality Assessment Using Multi-Pooled, Saliency Weighted Deep Features and Decision Fusion
title_fullStr No-Reference Video Quality Assessment Using Multi-Pooled, Saliency Weighted Deep Features and Decision Fusion
title_full_unstemmed No-Reference Video Quality Assessment Using Multi-Pooled, Saliency Weighted Deep Features and Decision Fusion
title_short No-Reference Video Quality Assessment Using Multi-Pooled, Saliency Weighted Deep Features and Decision Fusion
title_sort no reference video quality assessment using multi pooled saliency weighted deep features and decision fusion
topic no-reference video quality assessment
convolutional neural network
decision fusion
url https://www.mdpi.com/1424-8220/22/6/2209
work_keys_str_mv AT domonkosvarga noreferencevideoqualityassessmentusingmultipooledsaliencyweighteddeepfeaturesanddecisionfusion