Recurrent neural networks for solving second-order cone programs

Chun-Hsu Ko Jein-Shan Chen Ching-Yu Yang

Machine Learning mathscidoc:1910.43881

Neurocomputing, 74, (17), 3646-3653, 2011.10
This paper proposes using the neural networks to efficiently solve the second-order cone programs (SOCP). To establish the neural networks, the SOCP is first reformulated as a second-order cone complementarity problem (SOCCP) with the KarushKuhnTucker conditions of the SOCP. The SOCCP functions, which transform the SOCCP into a set of nonlinear equations, are then utilized to design the neural networks. We propose two kinds of neural networks with the different SOCCP functions. The first neural network uses the FischerBurmeister function to achieve an unconstrained minimization with a merit function. We show that the merit function is a Lyapunov function and this neural network is asymptotically stable. The second neural network utilizes the natural residual function with the cone projection function to achieve low computation complexity. It is shown to be Lyapunov stable and converges globally
No keywords uploaded!
[ Download ] [ 2019-10-20 22:35:57 uploaded by Jein_Shan_Chen ] [ 214 downloads ] [ 0 comments ]
  title={Recurrent neural networks for solving second-order cone programs},
  author={Chun-Hsu Ko, Jein-Shan Chen, and Ching-Yu Yang},
Chun-Hsu Ko, Jein-Shan Chen, and Ching-Yu Yang. Recurrent neural networks for solving second-order cone programs. 2011. Vol. 74. In Neurocomputing. pp.3646-3653.
Please log in for comment!
Contact us: | Copyright Reserved