Training algorithm matters for the performance of neural network potential: A case study of Adam and the Kalman filter optimizers
One hidden yet important issue for developing neural network potentials (NNPs) is the choice of training algorithm. In this article, we compare the performance of two popular training algorithms, the adaptive moment estimation algorithm (Adam) and the extended Kalman filter algorithm (EKF), using th...
Saved in:
| Published in | The Journal of chemical physics Vol. 155; no. 20; pp. 204108 - 204116 |
|---|---|
| Main Authors | , , , |
| Format | Journal Article |
| Language | English |
| Published |
Melville
American Institute of Physics
28.11.2021
|
| Subjects | |
| Online Access | Get full text |
| ISSN | 0021-9606 1089-7690 1520-9032 1089-7690 |
| DOI | 10.1063/5.0070931 |
Cover
| Summary: | One hidden yet important issue for developing neural network potentials (NNPs) is the choice of training algorithm. In this article, we compare the performance of two popular training algorithms, the adaptive moment estimation algorithm (Adam) and the extended Kalman filter algorithm (EKF), using the Behler–Parrinello neural network and two publicly accessible datasets of liquid water [Morawietz et al., Proc. Natl. Acad. Sci. U. S. A. 113, 8368–8373, (2016) and Cheng et al., Proc. Natl. Acad. Sci. U. S. A. 116, 1110–1115, (2019)]. This is achieved by implementing EKF in TensorFlow. It is found that NNPs trained with EKF are more transferable and less sensitive to the value of the learning rate, as compared to Adam. In both cases, error metrics of the validation set do not always serve as a good indicator for the actual performance of NNPs. Instead, we show that their performance correlates well with a Fisher information based similarity measure. |
|---|---|
| Bibliography: | ObjectType-Article-1 SourceType-Scholarly Journals-1 ObjectType-Feature-2 content type line 14 content type line 23 |
| ISSN: | 0021-9606 1089-7690 1520-9032 1089-7690 |
| DOI: | 10.1063/5.0070931 |