Translator Disclaimer
6 April 1995 Boundary variance reduction for improved classification through hybrid networks
Author Affiliations +
Several researchers have experimentally shown that substantial improvements can be obtained in difficult pattern recognition problems by combining or integrating the outputs of multiple classifiers. This paper provides an analytical framework that quantifies the improvements in classification results due to linear combining. We show that combining networks in output space reduces the variance of the actual decision region boundaries around the optimum boundary. In the absence of network bias, the added classification error is directly proportional to the boundary variance. Moreover, if the network errors are independent, then the reduction in variance boundary location is by a factor of N, the number of classifiers that are combined. In the presence of network bias, the reductions are less than or equal to N, depending on the interaction between network biases. We discuss how the individual networks can be selected to achieve significant gains through combining, and support them with experimental results on 25-dimensional sonar data. The analysis presented here facilitates the understanding of the relationships among error rates, classifier boundary distributions, and combining in output space.
© (1995) COPYRIGHT Society of Photo-Optical Instrumentation Engineers (SPIE). Downloading of the abstract is permitted for personal use only.
Kagan Tumer and Joydeep Ghosh "Boundary variance reduction for improved classification through hybrid networks", Proc. SPIE 2492, Applications and Science of Artificial Neural Networks, (6 April 1995);

Back to Top