An advanced conjugate gradient training algorithm based on a modified secant equation Print

I.E. Livieris and P. Pintelas, An Advanced Conjugate Gradient Training Algorithm Based on a Modified Secant Equation, ISRN Artificial Intelligence, 2012.


Abstract - Conjugate gradient methods constitute excellent neural network training methods, characterized by their simplicity, numerical efficiency and their very low memory requirements. In this paper, we propose a conjugate gradient neural network training algorithm which guarantees sufficient descent using any line search, avoiding thereby the usually inefficient restarts. Moreover, it achieves a high-order accuracy in approximating the second order curvature information of the error surface by utilizing the modified secant condition proposed by Li et al. (J. Comput. Appl. Math. 202(2):523--539, 2007). Under mild conditions, we establish that the proposed method is globally convergent for general functions under the strong Wolfe conditions. Experimental results provide evidence that our proposed method is preferable and in general superior to the classical conjugate gradient methods has a potential to significantly enhance the computational efficiency and robustness of the training process.