2 March 1994 Application of backpropagation neural architectures to the realization of control transfer functions and compensators
Author Affiliations +
A method is developed to design simulations of neural-network based transfer functions, applicable to both linear and nonlinear structures. The algorithm used to implement the trainable neural mechanism is backpropagation. Using the trained structures as building blocks, a neural architecture is constructed in order to drive systems from expected inputs to satisfactory transient and steady-state output performance, in effect, the scope of control compensation; this method results in the design of neural-net control compensators. The algorithms are coded in a PC-based prolog, traditionally used for rule-based logic and Artificial Intelligence, rather than for Neural or Fuzzy models. Given a sequence representing the time-sample of a desired control input trajectory that will drive the plant to a desired output response, such a control input will be modelled as the desired output layer of an antecedent network driven by an error vector consistent with the closed-loop system's commanded behavior. This Controller network is trained to provide such an output profile for all expected inputs, in accordance with arbitrary specifications of rise-time, permitted overshoot, settling time, etc. The control vectors are generated as a by-product of this training. Additionally, a correlation is investigated between classical control parameters and the characteristics of the weight matrices, threshold vectors, and representation traits of the converged neural nets.
© (1994) COPYRIGHT Society of Photo-Optical Instrumentation Engineers (SPIE). Downloading of the abstract is permitted for personal use only.
Regino R. Diaz-Robainas, Regino R. Diaz-Robainas, Abhijit S. Pandya, Abhijit S. Pandya, Ming Z. Huang, Ming Z. Huang, } "Application of backpropagation neural architectures to the realization of control transfer functions and compensators", Proc. SPIE 2243, Applications of Artificial Neural Networks V, (2 March 1994); doi: 10.1117/12.169985; https://doi.org/10.1117/12.169985


Back to Top