Weight Perturbation: An Optimal Architecture and Learning Technique for Analog VLSI Feedforward and Recurrent Multilayer Networks

Previous work on analog VLSI implementation of multilayer perceptrons with on-chip learning has mainly targeted the implementation of algorithms such as back-propagation. Although back-propagation is efficient, its implementation in analog VLSI requires excessive computational hardware. It is shown that using gradient descent with direct approximation of the gradient instead of back-propagation is more economical for parallel analog implementations. It is shown that this technique (which is called ;weight perturbation') is suitable for multilayer recurrent networks as well. A discrete level analog implementation showing the training of an XOR network as an example is presented.

[1]  Ra Reis A Topological Evaluator as the First Step in VLSI Design , 1982 .

[2]  Gerhard Zimmerman,et al.  A new area and shape function estimation technique for VLSI layouts , 1988, 25th ACM/IEEE, Design Automation Conference.Proceedings 1988..

[3]  Xinghao Chen,et al.  A module area estimator for VLSI layout , 1988, 25th ACM/IEEE, Design Automation Conference.Proceedings 1988..

[4]  Marwan A. Jabri,et al.  PIAF: A Knowledge-based/Algorithmic Top-down Floorplanning System , 1989, 26th ACM/IEEE Design Automation Conference.

[5]  Jenq-Neng Hwang,et al.  Parallel algorithms/architectures for neural networks , 1989, J. VLSI Signal Process..

[6]  Marwan A. Jabri,et al.  PIAF: efficient IC floor planning , 1989, IEEE Expert.

[7]  Fernando J. Pineda,et al.  Recurrent Backpropagation and the Dynamical Approach to Adaptive Neural Computation , 1989, Neural Computation.

[8]  Fadi J. Kurdahi,et al.  Techniques for area estimation of VLSI layouts , 1989, IEEE Trans. Comput. Aided Des. Integr. Circuits Syst..

[9]  Bernard Widrow,et al.  30 years of adaptive neural networks: perceptron, Madaline, and backpropagation , 1990, Proc. IEEE.

[10]  Francesco Piazza,et al.  Fast neural networks without multipliers , 1993, IEEE Trans. Neural Networks.