11 March 2015 Video quality assessment via gradient magnitude similarity deviation of spatial and spatiotemporal slices
Author Affiliations +
Video quality assessment (VQA) has been a hot topic due to the rapidly increasing demands in related video applications. The existing state-of-art full reference (FR) VQA metric ViS3 uses adapted the Most Apparent Distortion (MAD) algorithm to capture spatial distortion first, and then quantifies the spatiotemporal distortion by spatiotemporal correlation and a HVS-based model from the spatiotemporal slices (STS) images. In this paper we argue that the STS images can provide enough information for measuring video distortion. Taking advantage of an effective and easy-applied FR image quality model GMSD, we propose to measure video quality by analysing the structural changes between the STS images of the reference videos and their distorted counterparts. This new VQA model is denoted as STS-GMSD. To further investigate the influence spatial dissimilarity, we also combine the frame-by-frame spatial GMSD factor with the STS-GMSD and propose another VQA model, named SSTS-GMSD. Extensive experimental evaluations on two benchmark video quality databases demonstrate that the proposed STS-GMSD outperforms the existing state-of-the-art FR-VQA methods. While STS-GMSD works all square with SSTS-GMSD, which validates that STS images contain enough information for FR-VQA model design.
© (2015) COPYRIGHT Society of Photo-Optical Instrumentation Engineers (SPIE). Downloading of the abstract is permitted for personal use only.
Peng Yan, Peng Yan, Xuanqin Mou, Xuanqin Mou, Wufeng Xue, Wufeng Xue, "Video quality assessment via gradient magnitude similarity deviation of spatial and spatiotemporal slices", Proc. SPIE 9411, Mobile Devices and Multimedia: Enabling Technologies, Algorithms, and Applications 2015, 94110M (11 March 2015); doi: 10.1117/12.2083283; https://doi.org/10.1117/12.2083283


Perceptual tools for quality-aware video networks
Proceedings of SPIE (February 02 2014)

Back to Top