Training Artificial Neural Networks: Backpropagation via Nonlinear Optimization
Abstract
In this paper we explore different strategies to guide backpropagation algorithm used for training artificial neural networks. Two different variants of steepest descent-based backpropagation algorithm, and four different variants of conjugate gradient algorithm are tested. The variants differ whether or not the time component is used, and whether or not additional gradient information is utilized during one-dimensional optimization. Testing is performed on randomly generated data as well as on some benchmark data regarding energy prediction. Based on our test results, it appears that the most promissing backpropagation strategy is to initially use steepest descent algorithm, and then continue with conjugate gradient algorithm. The backpropagation through time strategy combined with conjugate gradients appears to be promissing as well.
Full Text:
PDFDOI: https://doi.org/10.2498/cit.2001.01.01
This work is licensed under a Creative Commons Attribution-NoDerivatives 4.0 International License.