Optimization schemes for neural network training

Neural networks are parameterized by a set of synaptic weights. The task of an optimization scheme for a neural network is to find a set of synaptic weights that make the network perform the desired function. The backpropagation learning method, quasi-Newton method, non-derivative quasi-Newton method, Gauss-Newton method, secant method and simulated Cauchy annealing method have been investigated. According to the computation time, convergence speed, and mean-squared error between the network outputs and desired results, the comparison of these learning methods has been presented. For learning a sine function, the quasi-Newton method can yield the best performance and the Gauss-Newton method also can provide a good promising result.<<ETX>>