Scanpath Prediction Via Semantic Representation of The Scene
Kepei Zhang, Meiqi Lu, Zheng Lu, Xuetao Zhang
-
SPS
IEEE Members: $11.00
Non-members: $15.00Length: 00:05:46
User-generated contents (UGC) have gained increased attention in the video quality community recently. Perceptual video quality assessment (VQA) of UGC videos is of great significance for content providers to monitor, process, and deliver massive numbers of UGC videos. Blind video quality prediction of UGC videos is challenging since complex mixtures of spatial and temporal distortions contribute to the overall perceptual quality. in this paper, we develop a simple, effective, and efficient BVQA framework (STS-QA) based on the statistical analysis of space-time slices (STS) of videos. Specifically, we extract spatio-temporal statistical features along different orientations of video STS, that capture directional global motion, then train a shallow quality predictor. The proposed framework can be used to easily extend any existing video/image quality model to account for temporal or motion regularities. Our experimental results on three publicly available UGC databases demonstrate that our proposed STS-QA model can significantly boost prediction performance compared to baselines. The code will be released at: \url{https://github.com/uniqzheng/STS_BVQA}.