3 June 1997 Modeling images and textures by minimax entropy
Author Affiliations +
Abstract
This article proposes a general theory and methodology, called the minimax entropy principle, for building statistical models for images (or signals) in a variety of applications. This principle consists of two parts. (1) Maximum entropy principle for feature binding (or fusion): for a given set of observed feature statistics, a distribution can be built to bind these feature statistics together by maximizing the entropy over all distributions that reproduce these feature statistics. The second part is the minimum entropy principle for feature selection: among all plausible sets of feature statistics, we choose the set whose maximum entropy distribution has the minimum entropy. Computational and inferential issues in both parts are addressed. The minimax entropy principle is then corrected by considering the sample variation in the observed feature statistics, and a novel information criterion is derived for feature selection. The minimax entropy principle is applied to texture modeling. Relationship between our theory and the mechanisms of neural computation is also discussed.
© (1997) COPYRIGHT Society of Photo-Optical Instrumentation Engineers (SPIE). Downloading of the abstract is permitted for personal use only.
Song-Chun Zhu, Song-Chun Zhu, YingNian Wu, YingNian Wu, David Mumford, David Mumford, } "Modeling images and textures by minimax entropy", Proc. SPIE 3016, Human Vision and Electronic Imaging II, (3 June 1997); doi: 10.1117/12.274536; https://doi.org/10.1117/12.274536
PROCEEDINGS
14 PAGES


SHARE
Back to Top