Advances in the image processing field have brought new methods which are able to perform complex tasks robustly.
However, in order to meet constraints on functionality and reliability, imaging application developers often
design complex algorithms with many parameters which must be finely tuned for each particular environment.
The best approach for tuning these algorithms is to use an automatic training method, but the computational
cost of this kind of training method is prohibitive, making it inviable even in powerful machines. The same
problem arises when designing testing procedures. This work presents methods to train and test complex image
processing algorithms in parallel execution environments. The approach proposed in this work is to use existing
resources in offices or laboratories, rather than expensive clusters. These resources are typically non-dedicated,
heterogeneous and unreliable. The proposed methods have been designed to deal with all these issues. Two methods are proposed: intelligent training based on genetic algorithms and PVM, and a full factorial design based on grid computing which can be used for training or testing. These methods are capable of harnessing the available computational power resources, giving more work to more powerful machines, while taking its unreliable nature into account. Both methods have been tested using real applications.