28 September 2016 Prediction of visual saliency in video with deep CNNs
Author Affiliations +
Abstract
Prediction of visual saliency in images and video is a highly researched topic. Target applications include Quality assessment of multimedia services in mobile context, video compression techniques, recognition of objects in video streams, etc. In the framework of mobile and egocentric perspectives, visual saliency models cannot be founded only on bottom-up features, as suggested by feature integration theory. The central bias hypothesis, is not respected neither. In this case, the top-down component of human visual attention becomes prevalent. Visual saliency can be predicted on the basis of seen data. Deep Convolutional Neural Networks (CNN) have proven to be a powerful tool for prediction of salient areas in stills. In our work we also focus on sensitivity of human visual system to residual motion in a video. A Deep CNN architecture is designed, where we incorporate input primary maps as color values of pixels and magnitude of local residual motion. Complementary contrast maps allow for a slight increase of accuracy compared to the use of color and residual motion only. The experiments show that the choice of the input features for the Deep CNN depends on visual task:for th eintersts in dynamic content, the 4K model with residual motion is more efficient, and for object recognition in egocentric video the pure spatial input is more appropriate.
Conference Presentation
© (2016) COPYRIGHT Society of Photo-Optical Instrumentation Engineers (SPIE). Downloading of the abstract is permitted for personal use only.
Souad Chaabouni, Jenny Benois-Pineau, Ofer Hadar, "Prediction of visual saliency in video with deep CNNs", Proc. SPIE 9971, Applications of Digital Image Processing XXXIX, 99711Q (28 September 2016); doi: 10.1117/12.2238956; https://doi.org/10.1117/12.2238956
PROCEEDINGS
14 PAGES + PRESENTATION

SHARE
Back to Top