DocumentCode :
1980685
Title :
Improved linear programming neural networks
Author :
Maa, Chia-Yiu ; Shanblatt, Michael A.
Author_Institution :
Dept. of Electr. Eng., Michigan State Univ., East Lansing, MI, USA
fYear :
1989
fDate :
14-16 Aug 1989
Firstpage :
748
Abstract :
It is shown that from the viewpoint of optimization theory a proper form of the Tank-Hopfield network for linear programming may be considered as a means to fulfil the Kuhn-Tucker optimality conditions. Due to the nature of the network, however, the convergence state is not the exact solution but an approximation. To get a better approximate solution, a new network formulation is introduced. The convergence state can be made very close to the exact solution by sufficiently increasing the network parameter λ. The result of this work is not limited to linear programming, it can be applied directly to any nonlinear programming problem whose objective function and constraints are convex and differentiable. This work may also be used as a foundation for the application of artificial neural networks to more general optimization problems
Keywords :
linear programming; neural nets; nonlinear programming; Kuhn-Tucker optimality conditions; Tank-Hopfield network; artificial neural networks; convergence state; linear programming neural networks; network formulation; network parameter; nonlinear programming problem; objective function; optimization theory; Analytical models; Artificial neural networks; Cities and towns; Convergence; Linear programming; Neural networks; Neurons; Nonlinear circuits; Traveling salesman problems; Vectors;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Circuits and Systems, 1989., Proceedings of the 32nd Midwest Symposium on
Conference_Location :
Champaign, IL
Type :
conf
DOI :
10.1109/MWSCAS.1989.101963
Filename :
101963
Link To Document :
بازگشت