PSO-GA based hybrid with Adam Optimization for ANN training with application in Medical Diagnosis

This paper introduces a novel PSO-GA based hybrid training algorithm with Adam Optimization and contrasts performance with the generic Gradient Descent based Backpropagation algorithm with Adam Optimization for training Artificial Neural Networks. We aim to overcome the shortcomings of the tradition...

Full description

Saved in:
Bibliographic Details
Published inCognitive systems research Vol. 64; pp. 191 - 199
Main Authors Yadav, Rajesh K., Anubhav
Format Journal Article
LanguageEnglish
Published Elsevier B.V 01.12.2020
Subjects
Online AccessGet full text
ISSN1389-0417
1389-0417
DOI10.1016/j.cogsys.2020.08.011

Cover

More Information
Summary:This paper introduces a novel PSO-GA based hybrid training algorithm with Adam Optimization and contrasts performance with the generic Gradient Descent based Backpropagation algorithm with Adam Optimization for training Artificial Neural Networks. We aim to overcome the shortcomings of the traditional algorithm, such as slower convergence rate and frequent convergence to local minima, by employing the characteristics of evolutionary algorithms. PSO has a property of faster convergence rate, which can be exploited to account for the slower pace of convergence of the traditional BP (which is due to low values of gradients). In contrast, the integration with GA complements the drawback of convergence to local minima as GA, possesses the capability of efficient global search. So by this integration of these algorithms, we propose our new hybrid algorithm for training ANNs. We compare both the algorithms for the application of medical diagnosis. Results display that the proposed hybrid training algorithm, significantly outperforms the traditional training algorithm, by enhancing the accuracies of the ANNs with an increase of 20% in the average testing accuracy and 0.7% increase in the best testing accuracy.
ISSN:1389-0417
1389-0417
DOI:10.1016/j.cogsys.2020.08.011