21 September 2004 Unsupervised optimization of support vector machine parameters
Author Affiliations +
Abstract
Selection of the kernel parameters is critical to the performance of Support Vector Machines (SVMs), directly impacting the generalization and classification efficacy of the SVM. An automated procedure for parameter selection is clearly desirable given the intractable problem of exhaustive search methods. The authors' previous work in this area involved analyzing the SVM training data margin distributions for a Gaussian kernel in order to guide the kernel parameter selection process. The approach entailed several iterations of training the SVM in order to minimize the number of support vectors. Our continued investigation of unsupervised kernel parameter selection has led to a scheme employing selection of the parameters before training occurs. Statistical methods are applied to the Gram matrix to determine kernel optimization in an unsupervised fashion. This preprocessing framework removes the requirement for iterative SVM training. Empirical results will be presented for the "toy" checkerboard and quadboard problems.
© (2004) COPYRIGHT Society of Photo-Optical Instrumentation Engineers (SPIE). Downloading of the abstract is permitted for personal use only.
Mary L. Cassabaum, Mary L. Cassabaum, Donald E. Waagen, Donald E. Waagen, Jeffrey J. Rodriguez, Jeffrey J. Rodriguez, Harry A. Schmitt, Harry A. Schmitt, } "Unsupervised optimization of support vector machine parameters", Proc. SPIE 5426, Automatic Target Recognition XIV, (21 September 2004); doi: 10.1117/12.542422; https://doi.org/10.1117/12.542422
PROCEEDINGS
10 PAGES


SHARE
Back to Top