Translator Disclaimer
23 December 1997 Illumination-invariant video segmentation by hierarchical robust thresholding
Author Affiliations +
Proceedings Volume 3312, Storage and Retrieval for Image and Video Databases VI; (1997)
Event: Photonics West '98 Electronic Imaging, 1998, San Jose, CA, United States
Many methods for video segmentation rely upon the setting and tuning of thresholds for classifying interframe distances under various difference measures. An approach that has been used with some success has been to establish statistical measures for each new video and identify camera cuts as difference values far from the mean. For this type of strategy the mean and dispersion for some interframe distance measure must be calculated for each new video as a whole. Here we eliminate this statistical characterization step and at the same time allow for segmentation of streaming video by introducing a preprocessing step for illumination-invariance that concomitantly reduces input values to a uniform scale. The preprocessing step provides a solution to the problem that simple changes of illumination in a scene, such as an actor emerging from a shadow, can trigger a false positive transition, no matter whether intensity alone or chrominance is used in a distance measure. Our means of discounting lighting change for color constancy consists of the simple yet effective operation of normalizing each color channel to length 1 (when viewed as a long, length-N vector). We then reduce the dimensionality of color to two-dimensional chromaticity, with values which are in 0..1. Chromaticity histograms can be treated as images, and effectively low-pass filtered by wavelet-based reduction, followed by DCT and zonal coding. This results in an indexing scheme based on only 36 numbers, and lends itself to a binary search approach to transition detection. To this end we examine distributions for intra-clip and inter-clip distances separately, characterizing each using robust statistics, for temporal intervals from 32 frames to 1 frame by powers of 2. Then combining transition and non-transition distributions for each frame internal, we seek the valley between them, again robustly, for each threshold. Using the present method values of precision and recall are increased over previous methods. Moreover, illumination change produces very few false positives.
© (1997) COPYRIGHT Society of Photo-Optical Instrumentation Engineers (SPIE). Downloading of the abstract is permitted for personal use only.
Jie Wei, Mark S. Drew, and Ze-Nian Li "Illumination-invariant video segmentation by hierarchical robust thresholding", Proc. SPIE 3312, Storage and Retrieval for Image and Video Databases VI, (23 December 1997);


Real-time detecting and tracking ball with OpenCV and Kinect
Proceedings of SPIE (September 28 2016)
New directions in video coding
Proceedings of SPIE (June 23 2003)
Novel technique for automatic key frame computing
Proceedings of SPIE (January 15 1997)
Logical unit and scene detection: a comparative survey
Proceedings of SPIE (January 28 2008)

Back to Top