A new conjugate gradient algorithm for training neural networks based on a modified secant equation

Conjugate gradient methods have been established as excellent neural network training methods, due to the simplicity of their iteration, numerical efficiency and their low memory requirements. In this work, we propose a conjugate gradient neural network training algorithm which guarantees sufficient...

Full description

Saved in:
Bibliographic Details
Published inApplied mathematics and computation Vol. 221; pp. 491 - 502
Main Authors Livieris, Ioannis E., Pintelas, Panagiotis
Format Journal Article
LanguageEnglish
Published Elsevier Inc 15.09.2013
Subjects
Online AccessGet full text
ISSN0096-3003
1873-5649
DOI10.1016/j.amc.2013.06.101

Cover

More Information
Summary:Conjugate gradient methods have been established as excellent neural network training methods, due to the simplicity of their iteration, numerical efficiency and their low memory requirements. In this work, we propose a conjugate gradient neural network training algorithm which guarantees sufficient descent using any line search, avoiding thereby the usually inefficient restarts. Moreover, it approximates the second order curvature information of the error surface with a high-order accuracy by utilizing a new modified secant condition. Under mild conditions, we establish that the global convergence of our proposed method. Experimental results provide evidence that our proposed method is in general superior to the classical conjugate gradient training methods and has a potential to significantly enhance the computational efficiency and robustness of the training process.
Bibliography:ObjectType-Article-2
SourceType-Scholarly Journals-1
ObjectType-Feature-1
content type line 23
ISSN:0096-3003
1873-5649
DOI:10.1016/j.amc.2013.06.101